Knowledge Enhanced Masked Language Model for Stance Detection

NAACL 2021  ·  Kornraphop Kawintiranon, Lisa Singh ·

Detecting stance on Twitter is especially challenging because of the short length of each tweet, the continuous coinage of new terminology and hashtags, and the deviation of sentence structure from standard prose. Fine-tuned language models using large-scale in-domain data have been shown to be the new state-of-the-art for many NLP tasks, including stance detection. In this paper, we propose a novel BERT-based fine-tuning method that enhances the masked language model for stance detection. Instead of random token masking, we propose using a weighted log-odds-ratio to identify words with high stance distinguishability and then model an attention mechanism that focuses on these words. We show that our proposed approach outperforms the state of the art for stance detection on Twitter data about the 2020 US Presidential election.

PDF Abstract

Datasets


Introduced in the Paper:

Twitter Stance Election 2020

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Stance Detection (US Election 2020 - Biden) Twitter Stance Election 2020 KE-MLM Average F1 0.7577 # 1
Stance Detection (US Election 2020 - Trump) Twitter Stance Election 2020 KE-MLM Average F1 0.7877 # 1

Methods