Abstract
Prompt tuning has achieved great success in various sentence-level classification tasks by using elaborated label word mappings and prompt templates. However, for solving token-level classification tasks, e.g., named entity recognition (NER), previous research, which utilizes N-gram traversal for prompting all spans with all possible entity types, is time-consuming. To this end, we propose a novel prompt-based contrastive learning method for few-shot NER without template construction and label word mappings. First, we leverage external knowledge to initialize semantic anchors for each entity type. These anchors are simply appended with input sentence embeddings as template-free prompts (TFPs). Then, the prompts and sentence embeddings are in-context optimized with our proposed semantic-enhanced contrastive loss. Our proposed loss function enables contrastive learning in few-shot scenarios without requiring a significant number of negative samples. Moreover, it effectively addresses the issue of conventional contrastive learning, where negative instances with similar semantics are erroneously pushed apart in natural language processing (NLP)-related tasks. We examine our method in label extension (LE), domain-adaption (DA), and low-resource generalization evaluation tasks with six public datasets and different settings, achieving state-of-the-art (SOTA) results in most cases.
| Original language | English |
|---|---|
| Pages (from-to) | 18357-18369 |
| Number of pages | 13 |
| Journal | IEEE Transactions on Neural Networks and Learning Systems |
| Volume | 35 |
| Issue number | 12 |
| DOIs | |
| State | Published - 2024 |
Keywords
- Contrastive learning
- few-shot learning
- information extraction
- named entity recognition (NER)
- prompting
Fingerprint
Dive into the research topics of 'Template-Free Prompting for Few-Shot Named Entity Recognition via Semantic-Enhanced Contrastive Learning'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver