Academic Research - United States
The United States leads global XAI research, with major contributions from both academic institutions and industry labs. This means that U.S. institutions set the research agenda for the global XAI community. Historically, the modern XAI field originated from early work at Carnegie Mellon in the 1990s on rule extraction, but gained momentum when DARPA launched the Explainable AI program in 2017, investing $75 million over four years to develop interpretable AI systems. According to Scopus data, U.S. institutions authored approximately 35% of XAI publications in 2024—more than any other country—with particularly strong representation in foundational methods (LIME, SHAP, TCAV) and healthcare applications. Compared to European research that emphasizes formal verification, U.S. work tends to prioritize practical deployment and scalability. The research landscape is characterized by close collaboration between universities and technology companies, with many influential researchers holding joint appointments or moving between academia and industry.
Two distinct schools of thought have emerged in U.S. XAI research, representing fundamentally different philosophies about how to achieve interpretable AI. The post-hoc explanation school, centered at University of Washington and Microsoft Research, focuses on explaining existing black-box models through techniques like LIME and SHAP. In contrast, the inherently interpretable school, led by Duke University's Cynthia Rudin, argues that black-box models should be replaced with transparent alternatives in high-stakes settings. According to Rudin (2019), this distinction matters because post-hoc explanations can be unfaithful to actual model behavior, whereas inherently interpretable models guarantee explanation accuracy. This debate has shaped the field's development and is discussed in the Evaluation & Frameworks page.
| Research Approach | Key Institutions | Advantages | Limitations |
|---|---|---|---|
| Post-hoc Explanation | UW, Microsoft, Google | Works with any model; no accuracy sacrifice | Explanations may not reflect true model reasoning |
| Inherently Interpretable | Duke, Harvard | Guaranteed faithful explanations; easier debugging | May require more feature engineering; perception of lower accuracy |
| Concept-Based | Google DeepMind, MIT | Human-understandable concepts; bridges technical and domain expertise | Requires concept annotation; limited to concept vocabulary |
Duke Interpretable ML Lab
Institution: Duke University, Durham, NC
Principal Investigator: Cynthia Rudin [Scholar] [ORCID]
Focus: Inherently interpretable machine learning models that achieve accuracy competitive with black-box methods. Pioneered the argument that interpretable models should be preferred over post-hoc explanations for high-stakes decisions.
Key Contributions: Falling Rule Lists, Optimal Sparse Decision Trees, Certifiably Optimal Rule Lists (CORELS), Scalable Bayesian Rule Lists
UW NLP Group
Institution: University of Washington, Seattle, WA
Key Researchers: Marco Tulio Ribeiro [Scholar], Carlos Guestrin
Focus: Model-agnostic explanation methods that can explain any machine learning model. Developed LIME and Anchors, two of the most widely-used XAI techniques.
Key Contributions: LIME (Local Interpretable Model-agnostic Explanations), Anchors, SP-LIME for global explanations, model debugging with explanations
Harvard Data to Actionable Knowledge Lab
Institution: Harvard University, Cambridge, MA
Principal Investigator: Finale Doshi-Velez [Scholar]
Focus: Rigorous evaluation frameworks for interpretability, interpretable reinforcement learning, and healthcare applications of interpretable ML.
Key Contributions: Taxonomy of interpretability evaluation (Doshi-Velez & Kim, 2017), interpretable clinical decision support systems
MIT CSAIL
Institution: MIT Computer Science and Artificial Intelligence Laboratory, Cambridge, MA
Key Researchers: Tommi Jaakkola [Scholar], Regina Barzilay [Scholar]
Focus: Rationalization methods for NLP, attention mechanisms as explanations, interpretable deep learning for chemistry and medicine.
Key Contributions: Rationalizing neural predictions, extractive rationale generation, interpretable drug discovery models
UC Irvine NLP Group
Institution: University of California, Irvine
Principal Investigator: Sameer Singh [Scholar]
Focus: Co-developer of LIME and Anchors, NLP explanation methods, model debugging and error analysis.
Key Contributions: LIME, Anchors, CheckList for behavioral testing of NLP models
Academic Research - Europe
European XAI research is distinguished by its emphasis on formal verification, theoretical foundations, and regulatory compliance. This orientation reflects the European Union's proactive approach to AI governance, particularly through the GDPR's "right to explanation" and the 2024 AI Act. Consequently, European researchers have developed methods that provide mathematical guarantees about explanation properties rather than relying solely on empirical evaluation. Unlike U.S. research that often prioritizes practical deployment, European work tends to emphasize theoretical rigor. However, both traditions have strengths: U.S. methods are more widely adopted in industry, while European methods offer stronger formal guarantees. According to bibliometric analysis, European institutions produced 28% of XAI publications in 2024—less than the U.S. but more than Asia-Pacific—with Germany and the United Kingdom contributing approximately 40% of the European total.
A distinctive feature of European XAI research is the integration of causality and uncertainty quantification into explanation frameworks. Because the Berlin-Tübingen axis (Fraunhofer HHI, MPI, TU Berlin) has established deep expertise in both neural network theory and causal inference, European methods like Layer-wise Relevance Propagation (LRP) offer formal decomposition of neural network outputs that can be verified mathematically. Compared to LIME and SHAP which rely on perturbation-based approximations, LRP provides exact attribution through backpropagation, making it particularly suitable for applications requiring mathematical guarantees. According to Miller (2019), European research has also contributed significantly to understanding explanation from a social science perspective, emphasizing that effective explanations must align with human cognitive processes. This theoretical rigor complements the more empirically-driven approach dominant in U.S. research, as discussed in the Evaluation & Frameworks page.
Fraunhofer Heinrich Hertz Institute
Institution: Fraunhofer HHI, Berlin, Germany
Key Researchers: Klaus-Robert Müller [Scholar], Wojciech Samek [Scholar]
Focus: Layer-wise Relevance Propagation (LRP), neural network interpretability, XAI for medical imaging and neuroscience.
Key Contributions: LRP method, Deep Taylor Decomposition, PatternNet/PatternAttribution, iNNvestigate toolbox
University of Melbourne
Institution: University of Melbourne, Australia
Principal Investigator: Tim Miller [Scholar]
Focus: Social science foundations of explanation, human-centered XAI design, explanation as a social process.
Key Contributions: Seminal paper on social science of explanation (Miller, 2019), contrastive explanation theory
Oxford Machine Learning Research Group
Institution: University of Oxford, UK
Key Researchers: Yarin Gal [Scholar]
Focus: Uncertainty quantification in deep learning, Bayesian deep learning, understanding model confidence.
Key Contributions: Dropout as a Bayesian Approximation, uncertainty estimation for AI safety and interpretability
Max Planck Institute for Intelligent Systems
Institution: MPI Tübingen, Germany
Key Researchers: Bernhard Schölkopf [Scholar]
Focus: Causal inference for machine learning, causal explanations of model behavior, kernel methods for interpretability.
Key Contributions: Causal learning theory, causal representation learning, kernel-based interpretability methods
Edinburgh Napier CogBID Lab
Institution: Edinburgh Napier University, UK
Principal Investigator: Amir Hussain [Scholar]
Focus: Cognitive computing, trustworthy AI, brain-inspired AI, sentiment analysis with explainability. Founding Editor-in-Chief of Cognitive Computation journal.
Key Contributions: Sentic computing, cognitive big data analysis, XAI for healthcare applications
Academic Research - Asia
Asian XAI research has grown rapidly, contributing approximately 25% of global publications in 2024, with China, India, and Japan as primary contributors. This means that the Asia-Pacific region has emerged as a significant force in XAI development. Studies show that Asian research groups have particularly focused on XAI applications in healthcare, manufacturing, and smart city infrastructure, reflecting regional priorities in AI deployment. For instance, BITS Pilani has pioneered XAI for IoT edge computing, whereas Tsinghua has focused on large-scale NLP explanations. Research demonstrates that Asian contributions have increased by over 300% since 2019, outpacing growth in both North America and Europe. Because many Asian countries are simultaneously developing AI capabilities and regulatory frameworks, XAI research in the region effectively bridges fundamental methods with immediate application needs. In other words, Asian XAI research is characterized by its practical orientation toward deployment rather than theoretical exploration.
A notable characteristic of Asian XAI research is the integration of explainability with edge computing and IoT systems. Groups at BITS Pilani, Tsinghua, and NUS have developed lightweight explanation methods suitable for resource-constrained environments, addressing the challenge of deploying interpretable AI in mobile health applications and industrial IoT. This practical orientation has resulted in techniques optimized for real-time explanation generation. Alternatively, some Asian groups focus on domain-specific applications rather than general methods. On the other hand, the advantage of this approach is faster deployment cycles versus the longer research timelines typical of Western fundamental research. For application-specific requirements, see the Applications & Domains page.
BITS Pilani IoT Lab
Institution: Birla Institute of Technology and Science Pilani, India
Principal Investigator: Vinay Chamola [Scholar]
Focus: Trustworthy AI, XAI for IoT and edge computing, security applications of explainable machine learning.
Key Contributions: XAI surveys, explainability in drone networks and autonomous systems
KIIT University AI Lab
Institution: Kalinga Institute of Industrial Technology, India
Key Researchers: Vikas Hassija [Scholar]
Focus: XAI review and taxonomy, blockchain and AI security, privacy-preserving machine learning.
Key Contributions: Comprehensive XAI reviews, trustworthy AI frameworks
King Fahd University of Petroleum and Minerals
Institution: KFUPM, Saudi Arabia
Key Researchers: Mufti Mahmud [Scholar]
Focus: Brain informatics, XAI for healthcare and cognitive computing, responsible AI applications.
Key Contributions: XAI in brain-computer interfaces, cognitive computing for healthcare
Industry Research Labs
Industry research labs have become central to XAI development because major technology companies face both internal needs (debugging production models, improving user trust) and external pressures (regulatory compliance, liability concerns). Research demonstrates that Microsoft Research and Google DeepMind have produced some of the most widely-adopted XAI tools, including SHAP and TCAV, which are now integrated into production systems serving billions of users. Therefore, industry XAI research tends to prioritize scalability and practical deployment over theoretical elegance. In practice, this means that industry tools often outperform academic prototypes in computational efficiency. Together, these industry labs have collectively advanced the field from academic concepts to production-ready tools used by millions of developers worldwide.
The relationship between industry and academia in XAI is particularly close, with frequent researcher movement and collaborative projects. For example, Scott Lundberg developed SHAP while at the University of Washington before moving to Microsoft Research, while Been Kim's TCAV work bridges her DeepMind position with ongoing academic collaborations. This industry-academia synergy has accelerated the translation of XAI methods from research papers to production tools, as evidenced by the rapid adoption of techniques covered in the Techniques & Methods page.
Industry labs differ from academic groups in their emphasis on production deployment. Whereas academic research often prioritizes theoretical properties (as in Lundberg & Lee's SHAP paper), industry implementations focus on computational efficiency and scalability. For instance, Microsoft's TreeSHAP achieves polynomial-time exact Shapley values specifically for tree ensembles, enabling SHAP deployment on models with millions of predictions per day. Similarly, Google's Integrated Gradients method was designed with gradient computation efficiency in mind, making it practical for explaining large neural networks in production. This practical focus has made industry toolkits like InterpretML, AI Explainability 360, and Captum the standard infrastructure for XAI deployment.
| Organization | Key Researchers | Focus Areas |
|---|---|---|
| Google DeepMind | Been Kim [Scholar] | TCAV (Testing with Concept Activation Vectors), concept-based explanations, human-AI interaction |
| Microsoft Research | Scott Lundberg [Scholar], Rich Caruana [Scholar] | SHAP values, TreeSHAP, InterpretML toolkit, Explainable Boosting Machines |
| Google Research | Mukund Sundararajan [Scholar] | Integrated Gradients, axiomatic attribution methods |
| IBM Research | Amit Dhurandhar, Karthikeyan Ramamurthy | Contrastive explanations, AI Fairness 360, AI Explainability 360 toolkit |
| Meta AI | Yann LeCun's team | Attention visualization, transformer interpretability, self-supervised learning explanations |
| NVIDIA Research | Various | Visual explanations for autonomous vehicles, real-time XAI for edge deployment |
Industry Toolkits
- SHAP (Microsoft) - Shapley value-based explanations
- LIME (UW/UCI) - Local interpretable model-agnostic explanations
- InterpretML (Microsoft) - Unified interpretability toolkit
- AI Explainability 360 (IBM) - Comprehensive explanation library
- Captum (Meta) - PyTorch model interpretability
Government Programs
Government initiatives have provided crucial funding and strategic direction for XAI research, particularly in defense and healthcare applications where accountability is paramount. For example, the DARPA XAI program (2017-2021), with over $75 million in funding, catalyzed the modern XAI field by establishing systematic evaluation methodologies. This means that explanation quality could be rigorously measured for the first time. As a result, many current XAI evaluation frameworks trace their origins to DARPA-funded research. Specifically, the program's emphasis on user studies and psychological evaluation influenced subsequent academic research. In other words, DARPA shifted the field from purely technical metrics toward human-centered assessment. Compared to earlier ad-hoc approaches, this systematic methodology represented a significant advancement. The literature shows that DARPA-funded teams produced over 200 publications that collectively shaped modern XAI practices.
Government interest in XAI reflects the growing deployment of AI in consequential domains including national security, healthcare policy, and social services. Because algorithmic decisions in these domains can affect fundamental rights and resource allocation, agencies like NIST have developed risk-based frameworks that specify explainability requirements proportional to decision impact. The EU AI Act, which entered into force in 2024, codifies similar risk-tiered explainability mandates as discussed in the Applications page.
DARPA Explainable AI (XAI) Program
Agency: Defense Advanced Research Projects Agency, USA
Duration: 2017-2021
Focus: Developing machine learning techniques that produce more explainable models while maintaining high performance. Funded 11 research teams across academia and industry.
Key Outcomes: Psychological evaluation frameworks for XAI, domain-specific explanation requirements, user studies on explanation effectiveness
NIST AI Risk Management Framework
Agency: National Institute of Standards and Technology, USA
Focus: Developing standards and guidelines for trustworthy AI, including explainability requirements for different risk levels and application contexts.
Key Outputs: AI Risk Management Framework (AI RMF), explainability guidelines for government AI procurement
European Commission AI Act
Agency: European Commission
Focus: Regulatory framework requiring transparency and explainability for high-risk AI systems. Establishes legal requirements for AI explanations in healthcare, employment, law enforcement, and other critical domains.
Key Requirement: High-risk AI systems must be "sufficiently transparent to enable users to interpret the system's output and use it appropriately"
Paper Authors
The authors of "Interpreting Black-Box Models: A Review on Explainable Artificial Intelligence" represent institutions across multiple countries:
| Author | Affiliation | Research Area |
|---|---|---|
| Vikas Hassija [Scholar] | KIIT University, India | AI/ML, Blockchain, Privacy |
| Vinay Chamola [Scholar] | BITS Pilani, India | IoT, 5G, Healthcare AI |
| Atishay Mahapatra | BITS Pilani, India | Machine Learning, XAI |
| Adit Singal | BITS Pilani, India | Machine Learning |
| Divyansh Goel | BITS Pilani, India | Deep Learning |
| Kaoru Huang | Industry | Applied ML |
| Simone Scardapane [Scholar] | Sapienza University of Rome, Italy | Deep Learning, Neural Networks |
| Indro Spinelli | Sapienza University of Rome, Italy | Graph Neural Networks |
| Mufti Mahmud [Scholar] | KFUPM, Saudi Arabia | Brain Informatics, Cognitive Computing |
| Amir Hussain [Scholar] | Edinburgh Napier University, UK | Cognitive Computing, Trustworthy AI |
Key Journals
- Cognitive Computation (Springer) - Brain-inspired computing and XAI (Editor-in-Chief: Amir Hussain)
- Nature Machine Intelligence - High-impact AI interpretability research
- Artificial Intelligence (Elsevier) - Foundational AI and explanation theory
- Journal of Machine Learning Research - Open-access XAI methods
- ACM TIIS - Human-AI interaction and explanation interfaces
- ACM FAccT - Fairness, accountability, and transparency
- Frontiers in Artificial Intelligence - Open-access XAI research
Conference Venues
Major XAI research is also presented at top AI conferences:
- NeurIPS - Neural Information Processing Systems (SHAP, TCAV papers)
- ICML - International Conference on Machine Learning
- KDD - Knowledge Discovery and Data Mining (original LIME paper)
- AAAI - Association for the Advancement of AI (Anchors paper)
- FAccT - ACM Conference on Fairness, Accountability, and Transparency
- ICLR - International Conference on Learning Representations