5 found
Order:
  1.  14
    Combining Heterogeneous Classifiers for Word-Sense Disambiguation.Dan Klein, Christopher D. Manning & Kristina Toutanova - unknown
    This paper discusses ensembles of simple but heterogeneous classifiers for word-sense disambiguation, examining the Stanford-CS224N system entered in the SENSEVAL-2 English lexical sample task. First-order classifiers are combined by a second-order classifier, which variously uses majority voting, weighted voting, or a maximum entropy model. While individual first-order classifiers perform comparably to middle-scoring teams’ systems, the combination achieves high performance. We discuss trade-offs and empirical performance. Finally, we present an analysis of the combination, examining how ensemble performance depends on error independence (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  2.  9
    Feature Selection for a Rich HPSG Grammar Using Decision Trees.Christopher D. Manning & Kristina Toutanova - unknown
    This paper examines feature selection for log linear models over rich constraint-based grammar (HPSG) representations by building decision trees over features in corresponding probabilistic context free grammars (PCFGs). We show that single decision trees do not make optimal use of the available information; constructed ensembles of decision trees based on different feature subspaces show signifi- cant performance gains (14% parse selection error reduction). We compare the performance of the learned PCFG grammars and log linear models over the same features.
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark   1 citation  
  3.  20
    Learning Random Walk Models for Inducing Word Dependency Distributions.Christopher D. Manning & Kristina Toutanova - unknown
    Many NLP tasks rely on accurately estimating word dependency probabilities P(w1|w2), where the words w1 and w2 have a particular relationship (such as verb-object). Because of the sparseness of counts of such dependencies, smoothing and the ability to use multiple sources of knowledge are important challenges. For example, if the probability P(N |V ) of noun N being the subject of verb V is high, and V takes similar objects to V , and V is synonymous to V , then (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  4.  12
    Parse Selection on the Redwoods Corpus: 3rd Growth Results.Christopher D. Manning & Kristina Toutanova - unknown
    This report details experimental results of using stochastic disambiguation models for parsing sentences from the Redwoods treebank (Oepen et al., 2002). The goals of this paper are two-fold: (i) to report accuracy results on the more highly ambiguous latest version of the treebank, as compared to already published results achieved by the same stochastic models on a previous version of the corpus, and (ii) to present some newly developed models using features from the HPSG signs, as well as the MRS (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  5.  13
    Optimizing Local Probability Models for Statistical Parsing.Mark Mitchell, Christopher D. Manning & Kristina Toutanova - unknown
    This paper studies the properties and performance of models for estimating local probability distributions which are used as components of larger probabilistic systems — history-based generative parsing models. We report experimental results showing that memory-based learning outperforms many commonly used methods for this task (Witten-Bell, Jelinek-Mercer with fixed weights, decision trees, and log-linear models). However, we can connect these results with the commonly used general class of deleted interpolation models by showing that certain types of memory-based learning, including the kind (...)
    Direct download  
     
    Export citation  
     
    Bookmark