21.04.2013 Views

Eckhard Bick - VISL

Eckhard Bick - VISL

Eckhard Bick - VISL

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Even trigrams, however, are far from expressing real syntactic structure, and the<br />

lexical collocation knowledge expressed in Hidden MMs is diluted considerably by the<br />

fact that it is seen through a word class filter. While the lexicalisation problem to a<br />

certain degree also haunts rule based grammars, the syntactic structure problem is<br />

"unique" to probabilistic HMM grammars and resides in the "Markov assumption" that<br />

p(tn|t1 ... tn-1) = p(tn|tn-1) (for bigrams), or = p(tn|tn-1tn-2) (for trigrams). In generative<br />

grammar, syntactic structure is handled in an explicit way, and functions both as the<br />

traditional objective and as the main tool of disambiguation. In CG, finally, syntactic<br />

structure can be expressed, but results as a kind of by-product of sequential contextual<br />

disambiguation rules. Of course, it does matter what the objective of disambiguation is:<br />

in fact, as shown in chapter 3.7.3, two thirds of all morphological CG-rules make do<br />

without "global" rules, i.e. they could be expressed as statistical n-gram transitions<br />

(though even here, most rules use a larger-than-trigram window), while only 10-20% of<br />

syntactic CG-rules can manage without unbounded contexts.<br />

One proposed solution to the syntax problem in probabilistic systems has been to<br />

expand context-free grammars (CFGs) into probabilistic context-free grammars<br />

(PCFGs), where CFG-productions are assigned conditional probabilities on the nonterminal<br />

being expanded, and the probability for a given syntactic (sub)tree can be<br />

computed as the product of the probabilities of all productions involved. The two<br />

readings of the sentence 'Einstein lectures last.', for instance, can be described by the<br />

following mini-PCFG, consisting of CFG-rules weighted with - arbitrary - production<br />

probabilities:<br />

(A) (B)<br />

(2) Einstein lectures last.<br />

1. S -> NP VP (p = 0.5) S S<br />

2. VP -> v (p = 0.3)<br />

3. VP -> v adv (p = 0.2)<br />

4. NP -> n (p =0.4) NP VP NP VP<br />

5. NP -> n n (p =0.1)<br />

n n v n v adv<br />

The complex NP reading (Einstein @>N lectures @SUBJ> last @FMV) involves<br />

productions 1, 2 and 5, yielding a complex probability of 0.5 x 0.3 x 0.1 = 0.015, while<br />

the single noun reading (Einstein @SUBJ> lectures @FMV last @

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!