Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Harmful Communication: Detection of Toxic Language and Threats on Swedish
Mind Intelligence Lab, Uppsala, Sweden.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.ORCID iD: 0000-0002-3724-7504
Uppsala Universitet, Uppsala, Sweden.
Stockholm University, Faculty of Social Sciences, Department of Computer and Systems Sciences.
Show others and affiliations
Number of Authors: 52024 (English)In: 2023 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), Association for Computing Machinery (ACM) , 2024, p. 624-630Conference paper, Published paper (Refereed)
Abstract [en]

Harmful communication, such as toxic language and threats directed toward individuals or groups, is a common problem on most social media platforms and online spaces. While several approaches exist for detecting toxic language and threats in English, few attempts have detected such communication in Swedish. Thus, we used transfer learning and BERT to train two machine learning models: one that detects toxic language and one that detects threats in Swedish. We also examined the intersection between toxicity and threat. The models are trained on data from several different sources, with authentic social media posts and data translated from English. Our models perform well on test data with an F1-score above 0.94 for detecting toxic language and 0.86 for detecting threats. However, the models' performance decreases significantly when they are applied to new unseen social media data. Examining the intersection between toxic language and threats, we found that 20\% of the threats on social media are not toxic, which means that they would not be detected using only methods for detecting toxic language. Our finding highlights the difficulties with harmful language and the need to use different methods to detect different kinds of harmful language.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM) , 2024. p. 624-630
Keywords [en]
toxic language, hate speech, threats
National Category
Information Systems
Research subject
Computer and Systems Sciences
Identifiers
URN: urn:nbn:se:su:diva-228009DOI: 10.1145/3625007.3627597ISI: 001191293500099Scopus ID: 2-s2.0-85190627665ISBN: 979-8-4007-0409-3 (print)OAI: oai:DiVA.org:su-228009DiVA, id: diva2:1849588
Conference
ASONAM '23: International Conference on Advances in Social Networks Analysis and Mining, 6-9 november 2023, Kusadasi Turkiye.
Available from: 2024-04-08 Created: 2024-04-08 Last updated: 2024-11-14Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Lindén, Kevin

Search in DiVA

By author/editor
Kaati, LisaLindén, KevinMoshfegh, Arvin
By organisation
Department of Computer and Systems SciencesStockholm University
Information Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 42 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf