Site icon Harvard Technology Review

Explain Yourself. Current Trends in Model Explainability

blue and yellow phone modules

Photo by Dan Cristian Pădureț on Pexels.com

blue and yellow phone modules
Photo by Dan Cristian Pădureț on Pexels.com

As we build bigger and more powerful machine learning (ML) models, we also create increasingly darker black box models. If you were to ask the scientists behind ChatGPT and Dall-E 2 how a certain response or image was created, you’d receive some long explanation about how the model is trained on a massive dataset with billions of parameters that can be tuned. However this leaves much to be desired. What does each parameter mean? If the model returns something inaccurate, why did it do that and how can we go about fixing it?

These questions become even more pressing as these ML models become increasingly more pervasive in our lives. ML models are increasingly being used in the healthcare industry to assist doctors in making decisions, diagnostics, and resource optimization. In the financial services industry, ML models are already being used to assess credit risk, financial crime risk, and to improve pricing. In criminal justice, models are being used to analyze DNA and forecast crime.

It is clear that in this small subset of fields that ML models are at a crucial decision making junction. These models significantly improve the efficiency and effectiveness of decision making which in turn provides significant value to all stakeholders. However without the ability to explain these decisions and how we arrived at them, trust in these ML models is significantly hindered. Thus as we develop bigger and better models, we also need to develop better mechanisms to map out and explain the decision making process of a model. 

By advancing mechanisms to explain ML models, we can build more trust in the models we use and lower the risk of the significant costs involved with a bad decision. These explanation mechanisms will also increasingly encourage the use of models by addressing the regulatory requirements involved with deploying black box models. These mechanisms also have significant implications for improved monitoring and auditing for the right results. Through these mechanisms, we can also promote fair high quality models and mitigate model drift – especially as these models are scaled.

One Quick Note: Difference between Explainability and Interpretability

When we talk about developing mechanisms to better understand ML models, there are two dimensions in which we can base these mechanisms. The first is interpretability of an ML model. Interpretability is defined as a human’s ability to intuitively understand a model. Specifically it refers to the ability to observe cause and effect in a model which in turn presents how a model works and makes a decision. Put another way, interpretability is the why and how of a model when generating predictions. Examples of interpretable models include decision trees and linear regression models where features and their weights are specifically highlighted. Finally interpretability exists on a spectrum where highly interoperable models are usually less powerful than less interpretable ones (i.e. neural networks).

The other dimension is explainability which is the ability to explain the ML behavior in human terms. Specifically it is the extent to which inner workings of a model can be explained in human terms. It is useful to note that explainability and interpretability go hand in hand but can be mutually exclusive. For instance neural networks lack interpretability as we don’t know how each neuron and its weight contributes to the final decision. However, by plugging in inputs and observing the outputs, neural networks can be explained. For instance assume we have a neural network to distinguish between cats and dogs. Then assume that by using different subsets of pictures we determine that the model categorizes animals with whiskers as cats and otherwise as dogs. In this way, we explained how the model categorizes cats and dogs without needing to understand the inner workings. In summation, interpretability refers to the intuitive understanding of the why and how of a ML model’s decision whereas explainability refers to the ability to explain ML behavior in human terms. 

In this article, we will only focus on the current trends in explainability. Specifically we will look at the growing number of XAI techniques followed by an exploration of the growing players in the space. XAI techniques will be quickly introduced and described. Then we look at the various open source tools available for XAI. Next a section is devoted to shining a light on the XAI work at Fiddler AI through an interview with Lior Berry, Director of Engineering, and Krishnaram Kenthapadi, Chief AI Officer & Scientist. Then we end off with a quick overview of the other industry players in the space.

XAI Techniques

There are three overall levels of XAI. There is local model explainability which asks specific questions about the model’s decisions. Then there’s cohort model explainability which uses subsets of data to test accuracy with new and unseen data. This can weed out potential bias in the model. Finally there’s global model explainability which focuses on the highest impact features on model outcomes and decisions. Below I’ve outlined several techniques that tackle each of these levels to varying degrees.

Open Source Tools

The growth of XAI has led to a proliferation of open source tools that enable developers to easily incorporate explainability into their models. These tools include:

Fiddler AI: Explainability to Make Better and Fairer Models

Fiddler AI is a rapidly growing startup looking to build greater trust into AI. To build this greater trust, Fiddler incorporates various tools in analytics, model monitoring, fairness, and explainable AI. In this way, Fiddler tackles the entire ML lifecycle and works to build trust at every step of the way. Explainable AI plays a crucial role in building this trust and Fiddler has done a lot of remarkable work in providing state of the art XAI and working on the cutting edge of XAI.

I was fortunate enough to have the opportunity to talk with Lior Berry, head of engineering at Fiddler AI, about XAI. We talked about how Fiddler AI incorporates XAI into their product offerings. XAI at Fiddler takes a five layered approach. The first layer consists of model monitoring to protect against extensive drift and data violations. The second layer explores the features of the presented data and further examines the dominance of each feature. Then the third layer provides point explanations while also exploring what drove a certain model decision. The fourth layer allows the developer to do a What-If analysis of the model by testing different values. Finally the fifth layer tackles fairness and ensures that sub-populations are trained fairly and that the model does not perform poorly on certain populations. We also talked about Lior’s perspective on the XAI sector and its challenges. He noted that an expansion of regulation in the use of AI has encouraged companies to adopt more XAI techniques. Amongst the challenges that Lior mentioned, was the challenge of explanations for image ML models. Fiddler is currently exploring the use of a k-mean and binary search based technique to further build explainability for these models.

In addition to talking with Lior Berry, I also talked to the brilliant Krishnaram Kenthapadi. Krishnaram is the chief data scientist and chief AI officer at Fiddler. These two roles encompass managing the data science team, dealing with product as well as engineering, talking with customers, and exploring AI value adds for customers. During our conversation, Krishnaram noted how XAI was an extremely young yet rapidly growing field. This means that many of the problems that arise are new problems that Krishnaram often discovers through conversations with his customers. Then with this insight on the problems, Krishnaram is able to steer his team and researchers towards solving these problems. From there our conversation drifted to how explainability will continue to lag behind as long as models become exponentially more complex. Explainability is made even more difficult by the fact that it cannot take a one-size-fits-all approach due to the differing business and legal needs of customers. From there, we began a more technical discussion about some of the new XAI approaches that Fiddler was tackling. In regards to the challenges of monitoring and understanding large language models (LLMs), Krishnaram mentioned that they were exploring a distributions based approach. For image-based models, they were looking into the use of random noise to stress test models.

These conversations have revealed not only the speed of development and deployment of XAI but also the growing challenges in the space. Companies like Fiddler are working alongside researchers to tackle these challenges head on as they work to make AI more explainable and in turn more responsible.

Other Growing Players in XAI

In addition to Fiddler, there is a growing list of well funded startups and existing Big Tech firms looking to tackle the XAI space. These companies are:

As models become more complex, we will start to see massive shortcomings in understanding the decisions they make. These shortcomings will come to clash with the growing regulations about the use of AI that will start to require more deterministic explanations for decision making by AI. These two counteracting forces, if unresolved, will lead to bottlenecks in the development of AI and the deployment of AI to improve our daily lives. Explainable AI provides one highly probable solution to this conflict. By making opaque black box models more transparent, explainable AI encourages greater trust in our models by providing insights into their decisions. A lot of work has already been done in explainable AI but much more still remains. One day we will be able to ask an AI to make a decision for us and further follow up by asking it to “explain yourself.”

Author

Exit mobile version