Early results for named entity recognition with conditional random fields, feature induction and web-enhanced lexicons
2003Vol. 4, pp. 188–191
Citations Over TimeTop 1% of 2003 papers
Abstract
Models for many natural language tasks benefit from the flexibility to use overlapping, non-independent features. For example, the need for labeled data can be drastically reduced by taking advantage of domain knowledge in the form of word lists, part-of-speech tags, character n-grams, and capitalization patterns. While it is difficult to capture such inter-dependent features with a generative probabilistic model, conditionally-trained models, such as conditional maximum entropy models, handle them well. There has been significant work with such models for greedy sequence modeling in NLP (Ratnaparkhi, 1996; Borthwick et al., 1998).
Related Papers
- → Randomized Kernel Approach for Named Entity Recognition in Tamil(2015)12 cited
- → Biomedical Named Entity Recognition Based on Skip-Chain CRFS(2012)22 cited
- Chinese Word Segmentation and Named Entity Recognition Based on Conditional Random Fields Models(2006)
- Two Step Chinese Named Entity Recognition Based on Conditional Random Fields Models(2008)
- → Conditional Random Fields for Biomedical Named Entity Recognition Revisited(2020)