https://peyrardm.github.io
👉 Stricter validity criteria?
👉 Maybe interpretability is inherently underdetermined? and we can only get control and predictability but not "understanding"
This is a fascinating topic, and we keep investigating. If you're interested, come and chat at ICLR!
👉 Stricter validity criteria?
👉 Maybe interpretability is inherently underdetermined? and we can only get control and predictability but not "understanding"
This is a fascinating topic, and we keep investigating. If you're interested, come and chat at ICLR!
- Multiple explanatory algorithms exists
- Even for one algorithm, there are many localizations in the network
Identifiability problems remain across scenarios: changing levels of over-parametrization, progress in training, multi-tasks, model size.
- Multiple explanatory algorithms exists
- Even for one algorithm, there are many localizations in the network
Identifiability problems remain across scenarios: changing levels of over-parametrization, progress in training, multi-tasks, model size.
Why? It allows us to enumerate all possible explanations and see how many pass various MI testing criteria.
Why? It allows us to enumerate all possible explanations and see how many pass various MI testing criteria.
Interpretability is also an exercise in causal inference!
Interpretability is also an exercise in causal inference!
To do so, many causal manipulations are performed to validate an explanation. But what if (many) other, incompatible explanations also pass the causal tests?
To do so, many causal manipulations are performed to validate an explanation. But what if (many) other, incompatible explanations also pass the causal tests?