Home > Research > Publications & Outputs > fBERT: A Neural Transformer for Identifying Off...

Links

Text available via DOI:

View graph of relations

fBERT: A Neural Transformer for Identifying Offensive Content

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date7/11/2021
Host publicationFindings of the Association for Computational Linguistics: EMNLP 2021
Place of PublicationStroudsburg, PA
PublisherAssociation for Computational Linguistics
Pages1792-1798
Number of pages7
ISBN (electronic)9781955917100
<mark>Original language</mark>English
EventThe 2021 Conference on Empirical Methods in Natural Language Processing - Barceló Bávaro Convention Centre, Punta Cana, Dominican Republic
Duration: 7/11/202111/11/2021
https://2021.emnlp.org/

Conference

ConferenceThe 2021 Conference on Empirical Methods in Natural Language Processing
Abbreviated titleEMNLP 2021
Country/TerritoryDominican Republic
CityPunta Cana
Period7/11/2111/11/21
Internet address

Conference

ConferenceThe 2021 Conference on Empirical Methods in Natural Language Processing
Abbreviated titleEMNLP 2021
Country/TerritoryDominican Republic
CityPunta Cana
Period7/11/2111/11/21
Internet address

Abstract

Transformer-based models such as BERT, XLNET, and XLM-R have achieved state-of-the-art performance across various NLP tasks including the identification of offensive language and hate speech, an important problem in social media. In this paper, we present fBERT, a BERT model retrained on SOLID, the largest English offensive language identification corpus available with over 1.4 million offensive instances. We evaluate fBERT’s performance on identifying offensive content on multiple English datasets and we test several thresholds for selecting instances from SOLID. The fBERT model will be made freely available to the community.