Template-Free Prompting for Few-Shot Named Entity Recognition via Semantic-Enhanced Contrastive Learning

Research output: Contribution to journalArticlepeer-review

17 Scopus citations

Abstract

Prompt tuning has achieved great success in various sentence-level classification tasks by using elaborated label word mappings and prompt templates. However, for solving token-level classification tasks, e.g., named entity recognition (NER), previous research, which utilizes N-gram traversal for prompting all spans with all possible entity types, is time-consuming. To this end, we propose a novel prompt-based contrastive learning method for few-shot NER without template construction and label word mappings. First, we leverage external knowledge to initialize semantic anchors for each entity type. These anchors are simply appended with input sentence embeddings as template-free prompts (TFPs). Then, the prompts and sentence embeddings are in-context optimized with our proposed semantic-enhanced contrastive loss. Our proposed loss function enables contrastive learning in few-shot scenarios without requiring a significant number of negative samples. Moreover, it effectively addresses the issue of conventional contrastive learning, where negative instances with similar semantics are erroneously pushed apart in natural language processing (NLP)-related tasks. We examine our method in label extension (LE), domain-adaption (DA), and low-resource generalization evaluation tasks with six public datasets and different settings, achieving state-of-the-art (SOTA) results in most cases.

Original languageEnglish
Pages (from-to)18357-18369
Number of pages13
JournalIEEE Transactions on Neural Networks and Learning Systems
Volume35
Issue number12
DOIs
StatePublished - 2024

Keywords

  • Contrastive learning
  • few-shot learning
  • information extraction
  • named entity recognition (NER)
  • prompting

Fingerprint

Dive into the research topics of 'Template-Free Prompting for Few-Shot Named Entity Recognition via Semantic-Enhanced Contrastive Learning'. Together they form a unique fingerprint.

Cite this