alwaysaditi's picture
End of training
dc78b20 verified
the importance of supertagging for wide-coverage ccg parsing this paper describes the role of supertagging in a wide-coverage ccg parser which uses a log-linear model to select an analysis. the supertagger reduces the derivation space over which model estimation is performed, reducing the space required for discriminative training. it also dramatically increases the speed of the parser. we show that large increases in speed can be obtained by tightly integrating the supertagger with the ccg grammar and parser. this is the first work we are aware of to successfully integrate a supertagger with a full parser which uses an automatically extracted grammar. we also further reduce the derivation space using constraints on category combination. the result is an accurate wide-coverage ccg parser which is an order of magnitude faster than comparable systems for other linguistically motivated formalisms. our scores give an indication of how supertagging accuracy corresponds to overall dependency recovery. we describe two log-linear parsing models for ccg: a normal-form derivation model and a dependency model. the ccg parsing consists of two phases: first the supertagger assigns the most probable categories toeach word, and then the small number of combinatory rules, plus the type-changing and punctuation rules, are used with the cky algorithm to build a packed chart. we propose a method for integrating the supertagger with the parser: initially a small number of categories is assigned to each word, and more categories are requested if the parser can not find a spanning analysis.