Better Simulations for Validating Causal Discovery with the DAG-Adaptation of the Onion Method
Erich Kummerfeld
Research Assistant Professor
Institute for Health Informatics, University of Minnesota
1
Hi!
Some background about me:
2
Causal discovery experience
3
Talk structure
4
Some examples of applications
5
How do nursing homes benefit from an on-site APRN?
6
What are the causes and effects of PTSD symptoms in populations with PTSD diagnosis?
7
What is the causal explanation for comorbid INTD and AUD?
8
What are the mechanisms that relate brain and behavior variables, and ultimately AUD?
9
How do individuals differ in terms of what causes them to drink?
10
How are brain networks causally connected during rest?
11
How is brain connectivity different during psychosis?
12
How do conspiracy theory beliefs relate to vaccine intentions and attitudes?
13
How can we improve treatment for psychosis?
14
What brain connectivity changes does neuromodulation cause, and what brain connectivities cause relapse?
Some position papers led by domain scientists promoting Causal Discovery
15
A podcast?!
16
Summary of work: everything is ad hoc
17
In most projects
The graph is merely one
Of multiple stepping stones
To the primary finding
18
What do domain scientists think of CD?
My impressions of applied scientists and clinicians
19
Trust in AI?
20
Passing the Statistics Gate
21
Validating Causal Discovery
22
Decision makers need confidence
23
The current reality of CD validation
24
Quick shoutout to other methods
�
25
One direction: improve simulations
26
How do existing simulation methods do on these criteria?
27
How do existing simulation methods do on these criteria?
Existing simulations are insufficient for even comparing relative performance of methods.
They are nowhere near providing evidence for expecting good performance on real data!
28
Our idea: sample uniformly
All of those points (and more!) can be achieved by sampling uniformly from the space of covariance matrices
29
Shoutout to Bryan
Bryan Andrews basically did everything.
30
What does sampling uniformly look like?
31
emitter
chain
collider
Simulated model parameters
32
ZARX: NOTEARS papers. Tetrad: BOSS paper.
Parameter distributions�(edges and errors)
33
ZARX: NOTEARS papers. Tetrad: BOSS paper.
R2 sortability?
34
ZARX: NOTEARS papers. Tetrad: BOSS paper.
Definition of evaluation statistics
35
CD methods on DaO data
36
dLiNGAM used the same models but with non-Gaussian exponential error��With standard evaluations, DaO is a difficult test for most algorithms
And non-DaO simulations…
37
ZARX: NOTEARS papers. Tetrad: BOSS paper.
Some algorithms that do very poorly on DaO suddenly do extremely well on non-DaO simulations
Going forward (1)
38
Going forward (2)
39
Some limitations
40
Discussion Questions
41
Discussion Questions
42
Discussion Questions
43
Discussion Questions
44
45