Severity: Warning
Message: file_get_contents(https://...@pubfacts.com&api_key=b8daa3ad693db53b1410957c26c9a51b4908&a=1): Failed to open stream: HTTP request failed! HTTP/1.1 429 Too Many Requests
Filename: helpers/my_audit_helper.php
Line Number: 176
Backtrace:
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 176
Function: file_get_contents
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 250
Function: simplexml_load_file_from_url
File: /var/www/html/application/helpers/my_audit_helper.php
Line: 3122
Function: getPubMedXML
File: /var/www/html/application/controllers/Detail.php
Line: 575
Function: pubMedSearch_Global
File: /var/www/html/application/controllers/Detail.php
Line: 489
Function: pubMedGetRelatedKeyword
File: /var/www/html/index.php
Line: 316
Function: require_once
Recent works have introduced prompt learning for Event Argument Extraction (EAE) since prompt-based approaches transform downstream tasks into a more consistent format with the training task of Pre-trained Language Model (PLM). This helps bridge the gap between downstream tasks and model training. However, these previous works overlooked the complex number of events and their relationships within sentences. In order to address this issue, we propose Event Co-occurrences Prefix Event Argument Extraction (ECPEAE). ECPEAE utilizes the co-occurrences events prefixes module to incorporate template information corresponding to all events present in the current input as prefixes. These co-occurring event knowledge assist the model in handling complex event relationships. Additionally, to emphasize the template corresponding to the current event being extracted and enhance its constraint on the output format, we employ the present event bias module to integrate the template information into the calculation of attention at each layer of the model. Furthermore, we introduce an adjustable copy mechanism to overcome potential noise introduced by the additional information in the attention calculation at each layer. We validate our model using two widely used EAE datasets, ACE2005-EN and ERE-EN. Experimental results demonstrate that our ECPEAE model achieves state-of-the-art performance on both the ACE2005-EN dataset and the ERE dataset. Additionally, according to the results, our model also can be adapted to the low resource environment of different training sizes effectively.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1038/s41598-024-82883-w | DOI Listing |
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC11682221 | PMC |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!