21.04.2013 Views

Eckhard Bick - VISL

Eckhard Bick - VISL

Eckhard Bick - VISL

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

example is a grammar of one terminal and the rule (x) -> ((x)) which produces an<br />

infinite language of “sentences” with paired brackets. One of the very few examples<br />

from the domain of natural language is Swiss German that has a construction where<br />

word order in two verbal sections of a sentence is cross-dependent. But since even such<br />

examples can be circumvented for constructions of finite depth, by writing as many<br />

context free rules to cover the phenomenon [for the bracketing example, (x) -> ((x)),<br />

((x)) -> (((x))), (((x))) -> ((((x)))), ...], most generative parsers have been built around<br />

context free grammars.<br />

In most languages, morphological structure is more linear than syntactic structure, and<br />

therefore easier to describe in an FSM framework. Thus, the TWOL-systems<br />

(Koskenniemi, 1983) used to supply analyser-input for most Constraint Grammars,<br />

describe words as linear morpheme transitions, allowing for phonetically motivated<br />

surface level changes at morpheme borders. Thus, the word ‘unrecognisable’ would be<br />

analysed as ‘un_recognis(e)_able’. Here, the FSM contains transition paths from<br />

preverbal prefix to verbal root, and from verbal root to postverbal suffix, expressed as<br />

so-called alternation of sub-lexica. A surface-level rule removes the ‘e’ of ‘recognise’<br />

because of the clash with the ‘a’ of ‘-able’. All inflexion and most cases of derivation<br />

and compounding can be handled this way 96 .<br />

On a syntactic level, on the other hand, it is very hard to imagine a FSM capable of<br />

describing free natural language, though the technique has been explored in recent years<br />

by, for instance, Atro Voutilainen (1994:32ff).<br />

The generative grammars of the context free type used for syntactic parsing, try to<br />

achieve several objectives at the same time: They analyse sentences by generating<br />

sentences, and they disambiguate both word class and function by assigning structure.<br />

While this does not by itself pose unresolvable technical problems, the conceptual<br />

priorities of generative grammar do seem to make it less efficient in a parsing context,<br />

i.e. for identifying “partes orationis”, or “parts-of-speech” on a morphological and<br />

functional level:<br />

• 1. In generative grammar, there is a tradition of assigning low priority to broad<br />

lexicography, which can be explained by the fact that “toy lexica” are fine for<br />

generating sentences, while being unsatisfactory for research on parsing [free]<br />

sentences.<br />

• 2. The constituent structure approach creates its own, theory-specific ambiguity<br />

priorities, some of which, like the scope of postnominal attachment or some cases of<br />

96 One might, of course, argue, that ‘un-’ as a preverbal prefix is limited to transitivized denominal verbs, usually ending in<br />

‘-ize’/’-ise’ or ‘-ate’. For productive word composition one would then need a higher level (context free) rule to describe the<br />

interdependence of the causative suffix and the antonymous prefix.<br />

- 144 -

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!