On the Analysis of Optical Mapping Data - University of Wisconsin ...
On the Analysis of Optical Mapping Data - University of Wisconsin ...
On the Analysis of Optical Mapping Data - University of Wisconsin ...
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
30<br />
1.0<br />
0.8<br />
0.6<br />
π^<br />
0.4<br />
0.2<br />
0.0<br />
0 5 10 15 20 25<br />
Fragment length<br />
Figure 2.3 The non-parametric MLE <strong>of</strong> π based on <strong>the</strong> Grenander estimator. The smooth<br />
grey curve is based on a naive zero truncated kernel density estimate <strong>of</strong> h.<br />
<strong>the</strong> so called Grenander estimator (see van der Vaart, 1998, Chapter 24). To see this, let G<br />
be <strong>the</strong> (known) CDF <strong>of</strong> Y . This happens to be exponential in this case, but any monotone<br />
decreasing density is sufficient. Consider <strong>the</strong> quantities <strong>of</strong> interest in a scale transformed by<br />
G, i.e., Ỹ = G(Y ), ˜X = G(X) and<br />
)<br />
˜π (ỹ) = P<br />
(Z = 1|Ỹ = ỹ = π(G −1 (ỹ))<br />
Let ˜h be density <strong>of</strong> ˜X. Since Ỹ ∼ U (0, 1) with constant density, ˜h(˜x) ∝ ˜π(˜x). Since G and<br />
hence G −1 are monotone increasing transformations,<br />
π ↑ =⇒ ˜π = π ◦ G −1 ↑ =⇒ ˜h ↑<br />
Hence, <strong>the</strong> MLE <strong>of</strong> ˜h, based on ˜X i ’s, is given by <strong>the</strong> Grenander estimator. The MLE <strong>of</strong> ˜π is<br />
proportional to that <strong>of</strong> ˜h, with <strong>the</strong> constant <strong>of</strong> proportionality obtained from <strong>the</strong> fact that<br />
lim ey→1 ˜π(ỹ) = 1. The estimator is inconsistent at ỹ = 1, i.e. y = ∞, but that is not <strong>of</strong><br />
interest to us. The MLE <strong>of</strong> π in <strong>the</strong> original scale is given by ̂π = ̂˜π ◦ G.<br />
Parametric estimation: Parametric forms <strong>of</strong> π are naturally easier to work with in practice.<br />
Obtaining maximum likelihood estimates in that case is straightforward in principle;<br />
most <strong>of</strong> <strong>the</strong> difficulty arises in obtaining <strong>the</strong> normalizing constant<br />
K(π) =<br />
∫ ∞<br />
0<br />
π(t) g(t) dt