Label-Guided relation prototype generation for Continual Relation Extraction.

PeerJ Comput Sci

Faculty of Electrical Engineering and Computer Science, University of Maribor, Maribor, Slovenia.

Published: October 2024

AI Article Synopsis

  • Continual relation extraction (CRE) focuses on extracting relationships as new data continuously arrives, addressing issues such as catastrophic forgetting.
  • Some existing methods use memory replay by storing historical instances or averaging hidden representations, but they fail to fully utilize the semantic information in relation labels and are limited by memory constraints.
  • The proposed Label-Guided Relation Prototype Generation enhances label representations and employs a multi-head attention mechanism to create better relation prototypes, along with a new loss function to preserve knowledge from previous tasks, leading to competitive performance in CRE.

Article Abstract

Continual relation extraction (CRE) aims to extract relations towards the continuous and iterative arrival of new data. To address the problem of catastrophic forgetting, some existing research endeavors have focused on exploring memory replay methods by storing typical historical learned instances or embedding all observed relations as prototypes by averaging the hidden representation of samples and replaying them in the subsequent training process. However, this prototype generation method overlooks the rich semantic information within the label namespace and are also constrained by memory size, resulting in inadequate descriptions of relation semantics by relation prototypes. To this end, we introduce an approach termed Label-Guided Relation Prototype Generation. Initially, we enhance the representations of label embeddings through a technique named label knowledge infusion. Following that, we utilize the multi-head attention mechanism to form relation prototypes, allowing them to capture diverse aspects of typical instances. The embeddings of relation labels are utilized at this stage, leveraging their contained semantics. Additionally, we propose a feature-based distillation loss function called multi-similarity distillation, to ensure the model retains prior knowledge after learning new tasks. The experimental results indicate that our method has achieved competitive performance compared to the state-of-the-art baseline models in CRE.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11622975PMC
http://dx.doi.org/10.7717/peerj-cs.2327DOI Listing

Publication Analysis

Top Keywords

prototype generation
12
label-guided relation
8
relation prototype
8
continual relation
8
relation extraction
8
relation prototypes
8
relation
7
generation continual
4
extraction continual
4
extraction cre
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!