02.03.2013 Views

Target Discovery and Validation Reviews and Protocols

Target Discovery and Validation Reviews and Protocols

Target Discovery and Validation Reviews and Protocols

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Gene Networks 41<br />

Fig. 6. Directed acyclic graph.<br />

In gene network inference by Bayesian networks, a gene is regarded as a<br />

r<strong>and</strong>om variable <strong>and</strong> represented by a node in a dag, <strong>and</strong> the relationship<br />

between a gene <strong>and</strong> its direct parents is represented as a conditional probability<br />

Pr(X j |Pa(X j )). Boolean networks are a deterministic system that explains a<br />

state of a gene by the Boolean function of the states of its direct parents. On<br />

the contrary, the Bayesian networks use a probabilistic nature to model the<br />

dependencies among genes. Because gene expression data measured by<br />

microarrays contain various kinds of noise, the probabilistic formulation in the<br />

Bayesian networks is advocated to draw more reliable information from<br />

microarray data.<br />

The decomposition in Eq. 1 holds when the structure of the directed acyclic<br />

graph G is given. However, for gene network estimation problem from microarray<br />

data, many parts of the true structure of the gene network are still unknown, <strong>and</strong><br />

we need to estimate them from microarray data. This structural learning problem<br />

can be viewed as a statistical model selection problem (17) <strong>and</strong> can be<br />

solved by using an information criterion, such as the Akaike information criterion,<br />

the Bayesian information criterion, <strong>and</strong> so on. In this chapter, we describe<br />

a method to select a network structure based on the Bayes approach. In the<br />

Bayes statistics, the optimal graph can be selected as the maximizer of the<br />

posterior probability of the graph conditional on the data.<br />

Suppose we have n independent observations D = {x 1 ,…,x n } for {X 1 ,…,X p },<br />

where the ith observation x i = (x i1 ,…,x ip ) t is the p-dimensional vector. Here a t<br />

indicates the transpose of the vector a. In our case, the posterior probability of<br />

the graph can be written as follows:<br />

π post(G | D) = π prior (G)p(D | G)<br />

,<br />

p(D)<br />

where πprior (G) is the prior probability of the graph, p(D) is the normalizing<br />

constant given by ∑ π (G)p(D | G) <strong>and</strong> p(D | G) is the marginal likelihood<br />

G prior<br />

obtained by

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!