Anisotropic span embeddings and the negative impact of higher-order inference for coreference resolution: An empirical analysis
dc.citation.volume | First View | |
dc.contributor.author | Hou F | |
dc.contributor.author | Wang R | |
dc.contributor.author | Ng S-K | |
dc.contributor.author | Zhu F | |
dc.contributor.author | Witbrock M | |
dc.contributor.author | Cahan SF | |
dc.contributor.author | Chen L | |
dc.contributor.author | Jia X | |
dc.date.accessioned | 2024-10-07T22:40:48Z | |
dc.date.available | 2024-10-07T22:40:48Z | |
dc.date.issued | 2024-01-25 | |
dc.description.abstract | Coreference resolution is the task of identifying and clustering mentions that refer to the same entity in a document. Based on state-of-the-art deep learning approaches, end-to-end coreference resolution considers all spans as candidate mentions and tackles mention detection and coreference resolution simultaneously. Recently, researchers have attempted to incorporate document-level context using higher-order inference (HOI) to improve end-to-end coreference resolution. However, HOI methods have been shown to have marginal or even negative impact on coreference resolution. In this paper, we reveal the reasons for the negative impact of HOI coreference resolution. Contextualized representations (e.g., those produced by BERT) for building span embeddings have been shown to be highly anisotropic. We show that HOI actually increases and thus worsens the anisotropy of span embeddings and makes it difficult to distinguish between related but distinct entities (e.g., pilots and flight attendants). Instead of using HOI, we propose two methods, Less-Anisotropic Internal Representations (LAIR) and Data Augmentation with Document Synthesis and Mention Swap (DSMS), to learn less-anisotropic span embeddings for coreference resolution. LAIR uses a linear aggregation of the first layer and the topmost layer of contextualized embeddings. DSMS generates more diversified examples of related but distinct entities by synthesizing documents and by mention swapping. Our experiments show that less-anisotropic span embeddings improve the performance significantly (+2.8 F1 gain on the OntoNotes benchmark) reaching new state-of-the-art performance on the GAP dataset. | |
dc.description.confidential | false | |
dc.edition.edition | 2024 | |
dc.identifier.citation | Hou F, Wang R, Ng SK, Zhu F, Witbrock M, Cahan SF, Chen L, Jia X. (2024). Anisotropic span embeddings and the negative impact of higher-order inference for coreference resolution: An empirical analysis. Natural Language Engineering. First View. | |
dc.identifier.doi | 10.1017/S1351324924000019 | |
dc.identifier.eissn | 1469-8110 | |
dc.identifier.elements-type | journal-article | |
dc.identifier.issn | 1351-3249 | |
dc.identifier.uri | https://mro.massey.ac.nz/handle/10179/71617 | |
dc.language | English | |
dc.publisher | Cambridge University Press | |
dc.publisher.uri | https://www.cambridge.org/core/journals/natural-language-engineering/article/anisotropic-span-embeddings-and-the-negative-impact-of-higherorder-inference-for-coreference-resolution-an-empirical-analysis/E59F426F59F86445BD3A0B9EA24EBB4A | |
dc.relation.isPartOf | Natural Language Engineering | |
dc.rights | (c) 2024 The Author/s | |
dc.rights | CC BY 4.0 | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.subject | Coreference resolution | |
dc.subject | higher-order inference | |
dc.subject | anisotropic span embeddings | |
dc.subject | contextualized representations | |
dc.title | Anisotropic span embeddings and the negative impact of higher-order inference for coreference resolution: An empirical analysis | |
dc.type | Journal article | |
pubs.elements-id | 486356 | |
pubs.organisational-group | Other |