1 of 37

Dynamics and context specificity in biological networks

Oct 7th 2021

Original slides created by Prof. Sushmita Roy

BMI 826-23 Computational Network Biology�Fall 2021

Anthony Gitter

https://compnetbiocourse.discovery.wisc.edu

2 of 37

Topics in this section

  • Types of dynamic network models
  • Non-stationary dynamic Bayesian networks
  • Input-Output Hidden Markov Models
  • Multi-task learning of graphs

3 of 37

Goals for today

  • Overview of different models to capture dynamics and context-specificity
  • Non-stationary dynamic Bayesian networks

4 of 37

What do we mean by context?

  • We will define context broadly
  • Context can be time, developmental stage, tissue, cell type, organ, disease, strains/individuals, species, stimuli

Different individuals

Different cell types

Different species

5 of 37

Network dynamics and context specificity

  • What does modeling “dynamics” mean?
    • The activity of nodes change over time and we want to model how this happens
    • The network (structure or parameters) changes with time
      • Structure can change due to changes at the node or edge level
  • What models can be used for capturing dynamics in networks?
  • How do these models capture dynamics?
    • Node level
    • Edge level

6 of 37

How do dynamics and context-specificity improve network inference?

  • Multiple related contexts

  • Temporal ordering of contexts

7 of 37

How do dynamics and context-specificity improve network inference?

  • Multiple related contexts
    • Assume similar contexts have similar network structures
    • Pool information across contexts
  • Temporal ordering of contexts
    • Resolve ambiguous independence relationships
    • Regulator activated/inhibited before target

8 of 37

What is special about dynamic models?

  • Time points could be treated as contexts

  • Dynamic models are affected if time points are re-ordered

0h

1h

2h

4h

8h

G1

G2

G3

G4

G5

0h

8h

2h

1h

4h

G1

G2

G3

G4

G5

Expression matrix

Reordered expression matrix

9 of 37

Strategies for capturing context and dynamics in networks

  • Which network inference methods have we already seen in class that account for dynamics?

10 of 37

Strategies for capturing context and dynamics in networks

  • Skeleton network-based approaches
  • Input-Output Hidden Markov Models (next week)
  • Dynamic networks with temporal transition of edges
    • Time-varying networks (today)
  • Multi-task learning approaches (next week)
  • Time-lagged network models
    • Ordinary Differential Equations (ODEs)
    • Time-lagged dependency networks
    • Vector autoregression
    • Dynamic Bayesian Networks and non-stationary extensions

11 of 37

Skeleton network-based approaches

  • Assume a background/skeleton network that defines the universe of possible edges
  • The network changes because node activity levels change
  • Skeleton networks:

A

B

Gene C

Transcription factors

C

A

B

DNA

X

Y

X

Y

Transcriptional regulatory

Protein-protein interaction

12 of 37

Skeleton network-based approaches

  • Assume a background/skeleton network that defines the universe of possible edges
  • The network changes because node activity levels change

Kim et al. 2013 Briefings in Bioinformatics

13 of 37

Skeleton network-based approaches

  • Given
    • A fixed network of nodes (genes, proteins, metabolites)
    • Activity levels of network nodes in a set of contexts (e.g. tissues, time points)
  • Do
    • Find subset of edges that are active in each context

14 of 37

Active subnetworks

  • Subnetwork that exhibits an unexpectedly high level of expression change

  • Not all nodes in the subnetwork must be active

  • Supports multiple conditions but not time

  • Yeast galactose pathway knockout example

Ideker et al. 2002 Bioinformatics

15 of 37

Timing in skeleton networks

  • TimeXNet: order nodes in network according to their timing

  • Based on network flow

  • Temporal Pathway Synthesizer related but stronger constraints (Köksal et al. 2018 Cell Reports)

Patil and Nakai 2014 BMC Systems Biology

16 of 37

Time-varying networks

  • The network changes between time at the node and edge levels
  • Example approaches
    • TESLA (Ahmed & Xing 2009 PNAS)
    • KELLER (Song et al. 2009 Bioinformatics)
  • TESLA
    • Temporally smoothed l1-regularized logistic regression
    • Based on temporal Exponential Random Graph Models
    • Assumes binary node values
    • Imposes a regularization term to make the graphs change smoothly over time
    • Estimates the graph structure by solving a set of local regularized logistic regression problems

17 of 37

TESLA for Drosophila development

  • Summarize 4028 genes by 43 ontology groups
  • 66 time points: embryonic, larval, pupal, and adulthood
    • Arbeitman et al. 2002 Science
  • Observe smooth changes over time

Ahmed & Xing 2009 PNAS

18 of 37

Time-lagged network models

  • Regulator activity does not instantaneously influence target
  • Look at past activity of regulators to predict current expression of targets

  • Previously discussed
    • Dynamic Bayesian networks
    • Inferelator

19 of 37

Time-lagged network models

Lu et al. 2021 PLOS Comp Bio

20 of 37

ODE models in networks

  • Assume we are modeling a gene regulatory network
  • Let xi denote the expression level of the ith gene
  • ODE for the change in expression of ith gene is

  • This is often approximated by a finite difference approximation

Greenfield et al. 2013 Bioinformatics

Regulators of ith gene

Expression level of regulator p

Degradation rate

21 of 37

Dynamic Bayesian networks (DBNs)

  • Suppose we have a time course of node activity levels
  • We assume that the activity levels at time t+1 depends upon t
    • But this does not change with t
  • DBNs can be used to model how the system evolves over time
    • We may have a different network at the first time point

22 of 37

Non-stationary dynamic Bayesian networks (nsDBNs)

  • Standard DBN assumes that the dependencies between the previous time point (t-1) and the current time point (t) are always the same
    • Dependencies do not depend on t
    • Stationarity assumption temporal model
  • Non-stationary DBNs relax this assumption
    • Robinson and Hartemink 2010 Journal of Machine Learning Research
    • Dependency structure can depend on the time window

23 of 37

Non-stationary dynamic Bayesian networks (nsDBNs)

  • When may there be different time windows in a dynamic biological process?

24 of 37

nsDBN graph structure transitions

  • Suppose we have three time windows
  • nsDBNs require us to define the dependency structure in each time window

Adapted from Robinson & Hartemink 2010

X1

X3

X4

X2

t1

t2

Transition times

X1

X3

X3

X4

X1

X3

X4

X2

X1

X2

X4

X4

X1

X3

X4

X2

Edge set changes between time windows

G1

G2

G3

25 of 37

nsDBN graph structure transitions

  • N discrete time points
  • m epochs or time periods between transitions times ti
  • Graph structures G1,..., Gm for each epoch

Robinson & Hartemink 2010

Transition t1

Epoch 2

Time point 719

Simulated dataset

26 of 37

Posterior distribution over graph structures

  • For m epochs, we would like to find G1,.., Gm by optimizing their posterior distribution

Prior over m graphs; can be used to incorporate our prior knowledge of how the graphs transition

Describe in terms of edge changes at each transition

27 of 37

Plate notation for graphical models

  • Shaded variables are observed
  • Unshared variables are unobserved (latent)

Dynamic Bayesian network

Graph structure

Pseudocount

One for each of

the n genes

Expression for gene i

Parent set of gene i

Conditional probability distribution parameters for gene i

Robinson & Hartemink 2010

28 of 37

nsDBN: Known Number of Known Times of transitions

nsDBN: KNKT

Graph structure at epoch j

Pseudocount

One for each of

the n genes

Expression for gene i

Parent set of gene i

in epoch j

Conditional probability distribution parameters for gene i for parent set h

Prior on edge transitions

Transition

Number of transitions

  • Shaded variables are observed
  • Unshared variables are unobserved (latent)

29 of 37

Prior knowledge about the dynamic process

  • Is the Known Number of Known Times setting realistic?

  • What would a Bayesian do?

30 of 37

31 of 37

nsDBN: Unknown Number of Unknown Times of transitions

nsDBN: UNUT

Graph at epoch j

Pseudocount

One for each of

the n genes

Expression for gene i

Parent set of gene i

in epoch j

Conditional probability distribution parameters for gene i for parent set h

Prior on edge transitions

Transition

Number of transitions

Prior on number of transitions

  • Shaded variables are observed
  • Unshared variables are unobserved (latent)

32 of 37

Assumptions on graph structures

  • Maximum parents per variable pmax
  • Maximum edge changes per transition smax
  • Truncated geometric prior on changes si per transition

  • Truncated geometric prior on number of epochs m

  • Convenient form for log likelihood calculations

 

 

 

 

33 of 37

Metropolis-Hastings for inference

  • Similar to inference procedure we saw for stationary DBN

  • What new states or move sets are needed?

34 of 37

Metropolis-Hastings for inference

  • Move sets
    • Add edge to G1
    • Delete edge from G1
    • Add edge to Δgi
    • Delete edge from Δgi
    • Move edge from Δgi to Δgj
    • Shift ti
    • Merge Δgi and Δgi+1
    • Split Δgi
    • Create Δgi
    • Delete Δgi

KNKT, KNUT, UNUT

KNUT, UNUT

UNUT

35 of 37

Evaluating nsDBN: simulations

Robinson & Hartemink 2010

36 of 37

Evaluating nsDBN: Drosophila

  • Same dataset used by TESLA but only use 11 genes
  • A: Stationary directed network (Zhao et al. 2006)
  • B: Non-stationary undirected network (Guo et al. 2007)
  • C: nsDBN in KNKT setting (Robinson & Hartemink 2010)

37 of 37

Take away points

  • If we can assume contexts or times are related, don’t learn independent networks for each
  • Temporal ordering can reduce ambiguity in network inference

  • Stationary assumption unrealistic in many cases
  • Non-stationary DBN is a generalization
    • Suitable for unknown transitions
    • Requires sufficient time points
    • Limit complexity of graph structure for scalability