Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Anaphora and Coreference Resolution: A Review (1805.11824v1)

Published 30 May 2018 in cs.CL

Abstract: Entity resolution aims at resolving repeated references to an entity in a document and forms a core component of NLP research. This field possesses immense potential to improve the performance of other NLP fields like machine translation, sentiment analysis, paraphrase detection, summarization, etc. The area of entity resolution in NLP has seen proliferation of research in two separate sub-areas namely: anaphora resolution and coreference resolution. Through this review article, we aim at clarifying the scope of these two tasks in entity resolution. We also carry out a detailed analysis of the datasets, evaluation metrics and research methods that have been adopted to tackle this NLP problem. This survey is motivated with the aim of providing the reader with a clear understanding of what constitutes this NLP problem and the issues that require attention.

Anaphora and Coreference Resolution: A Review

The paper by Sukthanker et al. provides a comprehensive survey of the field of entity resolution in NLP, specifically focusing on the tasks of anaphora resolution (AR) and coreference resolution (CR). These tasks are crucial for various NLP applications, including machine translation, sentiment analysis, and summarization. The paper explores the nuances of AR and CR, offering insights into the methodologies, datasets, and evaluation metrics that have been employed to tackle these problems.

Entity resolution involves identifying and linking multiple references to the same entity within a text. Anaphora resolution typically addresses references that point back to previously mentioned items in a discourse, such as pronouns linking to noun phrases. Coreference resolution encompasses a broader range of phenomena, including cases where distinct expressions refer to the same real-world entity without strict adherence to discourse order. In contrast to AR, CR has a wider scope and is more challenging due to its extra-linguistic considerations.

Reference Types and Constraints

The paper systematically categorizes various types of references encountered in natural language, such as zero anaphora, one anaphora, demonstratives, presuppositions, and discontinuous sets. It highlights challenges like resolving zero anaphora, where antecedents are implicit, and bridging anaphora, which require inferencing unstated connections. It also discusses non-anaphoric pronominal references, which can mislead AR algorithms.

A thorough account of constraints applied in AR is provided, including gender and number agreement, person agreement, recency, and discourse structure. These constraints are essential for eliminating incorrect antecedents and narrowing down possible referents. However, the authors note that such constraints can be language-specific and are not universally applicable.

Datasets and Evaluation Metrics

The survey covers significant datasets used for AR and CR, such as MUC, ACE, and CoNLL, highlighting their differences in annotation schemes, languages, and domains. It stresses the importance of dataset choice in comparing the performance of AR and CR methods due to variations in their handling of reference types. The paper also reviews evaluation metrics like MUC, B-cubed, CEAF, and the CoNLL score, discussing their strengths and weaknesses in assessing CR performance.

Methodological Shifts

The review traces the evolution from rule-based approaches to machine learning and, more recently, deep learning models for AR and CR. Early methods relied heavily on syntactic and semantic rules, but these have largely been supplanted by statistical and machine learning techniques that leverage annotated data. Mention pair models, entity-mention models, and mention ranking models are explored, with a special focus on the latter's ability to capture competition among antecedents.

The integration of deep learning into CR has led to models that learn feature representations directly from data, reducing dependence on hand-engineered features. These models, including those employing LSTM and CNN architectures, have shown success in capturing complex linguistic phenomena and modeling global context. Nonetheless, the paper acknowledges challenges such as the high computational costs and complexity inherent in deep neural networks.

Application to Sentiment Analysis

The authors emphasize the relevance of AR and CR to sentiment analysis, particularly in aspect-based sentiment tasks where resolving references to product aspects is crucial. They illustrate the benefits of employing AR to develop a global understanding of sentiment and infer pronominal references across review corpora.

Outstanding Issues

The paper addresses ongoing debates and challenges in the field, such as the need for common sense knowledge integration and the development of more comprehensive evaluation metrics. It advocates for future datasets and models to be explicit about the types of references they tackle and suggests cross-domain evaluations to better understand the limitations of current approaches.

Conclusion

Sukthanker et al.'s review offers a robust groundwork for understanding the complexities and progress in AR and CR research. By elucidating the taxonomies, methodologies, and evolving trends in this domain, the paper provides valuable guidance for researchers aiming to contribute to or utilize entity resolution in NLP applications. The authors effectively highlight both the achievements and pending challenges in the field, setting the stage for continued advancement in AR and CR technologies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Rhea Sukthanker (3 papers)
  2. Soujanya Poria (138 papers)
  3. Erik Cambria (136 papers)
  4. Ramkumar Thirunavukarasu (1 paper)
Citations (160)