Mauro D. Ríos (The Internet Society) has posted “Can What an AI Produces be Understood and Unraveled” on SSRN. Here is the abstract:
Over the past few decades, AI has radically transformed industries as diverse as medicine and finance, providing solutions with high levels of efficiency and accuracy that were previously unattainable (Revolutionizing healthcare, 2023)
However, the sophistication of these models, which include deep neural networks with millions of parameters and sophisticated mechanisms for producing results, has led to the perception that they operate as a “black box” whose internal logic is inaccessible to human understanding (Hyperight, 2024)
Far from being an intrinsic feature of AI, this opacity is the result of both the volume and heterogeneity of training data and the lack of adequate methodologies to record and unravel each phase of the internal calculation (Stop Explaining Black Box Models, 2022).
To overcome these myths and reveal the “why” and “how” of AI decisions, various interpretability and auditing techniques have been developed. In addition, relevance propagation methodologies, such as Layer‐Wise Relevance Propagation (LRP), make it possible to track, layer by layer, the influence of each “digital neuron” on an AI’s final decision (Montavon et al., 2019).
While these tools offer an unprecedented level of visibility, their practical application involves addressing challenges of scale and computational cost. Exhaustive logging of execution traces and parameters during training demands distributed computing infrastructures and storage systems designed for metadata versioning (Unfooling Perturbation-Based Post Hoc Explainers, 2022).
A comprehensive understanding of AI processes requires not only the use of advanced interpretability techniques, but also the establishment of governance frameworks and structured documentation. Reports from organizations such as the Centre for International Governance Innovation (CIGI) underline the need for accountability policies that require detailed records of each phase of the AI model lifecycle, from data selection to production (Explainable AI Policy, 2023). Without these mechanisms, the aspiration to full interpretability will remain limited by practical and organizational barriers, not because we can claim that we do not know why an AI does what it does, but because we have failed to implement the appropriate mechanisms to know, and thus we will be compromising transparency and trust in critical AI applications.
Knowing what an AI does and why is then a possibility that is in our hands but requires instruments, time and resources that we must decide if they are justified in each case or we will be selective when.
