WhiteBox – explainable models for human and artificial intelligence

The LOEWE Research Cluster WhiteBox is aimed at developing methods at the intersection between Cognitive Science and AI to make human and artificial intelligence more understandable

Project Introduction

Until a few years ago, intelligent systems such as robots and digital voice assistants had to be tailored towards narrow and specific tasks and contexts. Such systems needed to be programmed and fine tuned by experts. But, recent developments in artificial intelligence have led to a paradigm shift: instead of explicitly representing knowledge about all information processing steps at time of development, machines are endowed with the ability to learn. With the help of machine learning it is possible to leverage large amounts of data samples, which hopefully transfer to new situations via pattern matching. Groundbreaking achievements in performance have been obtained over the last years with deep neural networks, whose functionality is inspired by the structure of the human brain. A large number of artificial neurons interconnected and organized in layers process input data under large computational costs. Although experts understand the inner working of such systems, as they have designed the learning algorithms, often they are not able to explain or predict the system’s intelligent behavior due to its complexity. Such systems end up as blackboxes raising the question of how such systems’ decisions can be understood and trusted.

Our basic hypothesis is that explaining an artificial intelligence system may not be fundamentally different from the task of explaining intelligent goal-directed behavior in humans. Behavior of a biological agent is also based on the information processing of a large number of neurons within brains and acquired experience. But, an explanation based on a complete wiring diagram of the brain and all its interactions with its environment may not provide an understandable explanation. Instead, explanations of intelligent behavior need to reside at a computationally more abstract level: they need to be cognitive explanations. Such explanations are developed in computational cognitive science. Thus, WhiteBox aims at transforming blackbox models into developing whitebox models through cognitive explanations that are interpretable and understandable.

Following our basic assumption, we will systematically develop and compare whitebox and blackbox models for artificial intelligence and human behavior. In order to quantify the differences between these models, we will not only develop novel blackbox and whitebox models, but also generate methods for the quantitative and interpretable comparison between these models. Particularly, we will develop new methodologies to generate explanations automatically by means of AI. As an example, deep blackbox models comprise deep neural networks whereas whitebox models can be probabilistic generative models with explicit and interpretable latent variables. Application of these techniques to intelligent goal directed human behavior will provide better computational explanations of human intelligent behavior as well as allow to transfer human level behavior to machines.



March 2022 – Meike Kietzmann joins the project – welcome to the team!

February 2022 – Asghar Mahmoudi Khomami joins the project – welcome to the team!

2021 – Ute Korn supports the project as an associated researcher since the end of 2021 – welcome to the team!

December 2021 – Due to the COVID-19 pandemic, the HWMK (Hessian State Ministry of Higher Education, Research and the Arts) extended the WhiteBox project duration until 31.12.2025.

November 2021 – Rabea Turon & Sven Schultze join the project – welcome to the team!

October 2021 – Two new introduction videos about project WhiteBox released in collaboration with ProLOEWE: Watch (1) Watch (2) (in German)

September 2021 – Rupert Mitchell joins the project – welcome to the team!

July 2021 – New video trailer released for the ProLOEWE Science Rallye!

July 2021 – WhiteBox starts a quiz video too in the frame of the ProLOEWE Science Rallye. The Rallye starts on August 2 on proloewe.de. Looking forward to puzzle enthusiasts!

July 2021 – Niteesh Midlagajni joins the project – welcome to the team!

June 2021 – Omid Mohseni Shorgheini joins the project – welcome to the team!

May 2021 – Maziar Sherbafi & Kai Ploeger join the project – welcome to the team!

January 2021 – Matthias Schultheis joins the project – welcome to the team!

January 2021 – Official start of project “WhiteBox – explainable models for human and artificial intelligence”

Project Details

  • Project: WhiteBox – explainable models for human and artificial intelligence (Erklärbare Modelle für menschliche und künstliche Intelligenz)
  • Project partners: Technical University of Darmstadt (TU Darmstadt)
  • Project duration: January 2021 – December 2025
  • Project funding: 4.7 Mio EUR
  • Funded by: Hessian State Ministry of Higher Education, Research and the Arts
  • Funding Line: LOEWE Research Cluster, Funding Round 13