Although neural networks have successfully revolutionised multiple domains, e.g., computer vision, they are still doubted in safety-critical applications, such as health diagnosis, credit allowance, or criminal justice, for they are still mostly not human-intelligible (Molnar 2020; Camburu 2020). Two approaches are typically used to make these models more plausible: 1) providing feature-based post-hoc explanations to a model and 2) creating self-explanatory models that generate natural language explanations (Camburu 2020). However, Rudin (2019) argues against a post-hoc justification for a black-box model, or a model that lacks a mechanistic basis. Instead, she postulates that there should be a so-called “Rashomon set,” which are a set of similarly accurate models containing at least one plausible model. Yet, any notion of Rashomon sets should be questioned in two ways: 1) can the Rashomon Set be realistically found? 2) and more importantly, can best fitting and interpretable models authentically explain the world we are living in?
An unproven, empirical theory
Intuitively, the notion of Rashomon sets is likely to be true. If the set is large enough, there is likely to be an accurate, plausible model, even if the possibility is low. However, it is not guaranteed that it is true in all cases. Firstly, the theory does not provide any convincing mechanistic explanation within itself. The reason Rudin makes to support her idea is mostly empirical, that is, it is always possible to find a both highly accurate and interpretable model in most scenarios she works with, ranging from New York City power grid to criminology data. However, there is no supporting theory to generalise the empirical evidence, and it is also technically difficult to prove or falsify such a theory, since we need to evaluate all known models in a Rashomon set in that case. Secondly, the theory is statistically probable, which means it should not be expected to be true for all scenarios. An analogue here is that although the water on Earth is highly likely to be saline statistically since saline water takes up over 97% of Earth’s water, yet undoubtedly there still exists freshwater (Shiklomanov 1993), where the high likelihood is not useful. Similarly, the theory may not help in specific scenarios, due to the domain-specific nature of interpretability; thus, we are unable to tell whether the model in a Rashomon set is plausible without a specific context (Martin 2019).
When human fails to interpret things
Practically, it is not always realistic to find a best-fitting interpretable model that outcompetes their black-box counterparts, due to limitations in data availability. For example, in genomic medicine, it is not until in recent years, we have come to take more omics approaches into considerations due to the complex mechanisms regulating phenotypic variation, which are beyond our expectation (Boyle et al. 2017). Yet, examining all epigenetic changes and other biological alterations is technically difficult, and even if possible, it will be too expensive, although it is helpful to build a biologically plausible model. Besides economic considerations, Li and Lehner (2020) prove that the phenotypic effect of two mutations combined is not accurately predictable, even when individual effects of a set of mutations and the mechanistic links are well-understood, due to biophysical ambiguity yet to be measured. Thus, chances are that not only do we fail to have enough data, but we may not fully realise what data we need to have, which reflects limitations of human knowledge that hinder the establishment of human-intelligible models, especially in poorly understood domains.
Why the Rashomon set is still important?
In the Japanese movie Rashomon, a murder is described by different witnesses in four contradictory ways (Davenport 2010). Consider all characters in the movie as functions, which receive the murder as an input, and produce different outputs which all make sense in a way. It is hard to guess the whole story without knowing the context related to each character. Similarly, in statistical modelling, Breiman (2001) coins the term “Rashomon effect,” where he postulates the multiplicity of good models, meaning that for a given dataset, it is possible to construct multiple well-performing models in term of their accuracy, despite their differences in internal structures and attendant explanations (Hancox-Li 2020). Furthermore, Rudin (2019) experimentally shows similar accuracy of popular machine learning models for a structured dataset, whereas complexity does not yield higher accuracy. Therefore, Rudin (2019) asserts that we should choose a more interpretable model, which should be more meaningful, reasonable and no less efficient.
Interpretability vs. Explainability: helpful in different ways
Rudin (2019) believes that interpretable models must be inherently interpretable, with a transparent internal structure, while an explainable model often contains two models, i.e., a black-box model and an explanatory model, which must contradict each other. Post-hoc explanations about these models must be at least partly wrong, or else the explanatory model should be used instead. Moreover, a meaningful model choice does not necessarily lead to an explainable model, especially when we don’t understand the meaning underlying the model. Often, our data often involve extraordinarily complex interactions between variables, which make simple, contrastive explanations that people would like to know highly unlikely. For example, in quantitative genetics, we could identify hundreds of loci responsible for a single trait but the contribution of even the most important loci is only moderate (Boyle et al. 2017), which makes the genetic explanation of a trait less intuitively straightforward. Therefore, Rudin also points out that her advocacy of interpretable models does not apply to complicated, less-structured, noisy data that lack representativeness, such as dataset used by computer vision, where too many different highly accurate models are unusual.
Ultimate consideration: who choose what we believe?
Furthermore, Hancox-Li (2020) and others argue against the robustness of the Rashomon set. A concrete example is made by Dong and Rudin (2020) who construct a Rashomon set for the recidivism data set used in the COMPAS algorithm. They find that lower importance of race is associated with higher importance of age or prior criminal history and vice versa, in the almost equally accurate predictive models given by the Rashomon set and that race is important only when age or prior criminal history are unimportant. However, although race is less important in some models in the set, it does not mean that it is objectively unimportant; Also, it is questionable how we should choose the explanations for the model. Although Rudin argues that a domain expert could help with model improvement, it is possible that a company deliberately hide all the other model choices, providing an explanation that is likely to be accepted, which may cause a moral crisis. Therefore, a description of variable-variable interaction will provide a fuller understanding of the real-world fact that is reflected by the Rashomon set (Hancox-Li 2020).