Argumentation and agreement: Annotating and evaluating Swedish corpora for argumentation mining

dc.contributor.authorLindahl, Anna
dc.date.accessioned2025-09-12T07:03:23Z
dc.date.available2025-09-12T07:03:23Z
dc.date.issued2025-09-12
dc.description.abstractArgumentation occurs in all parts of life, and as such, is studied across disciplines. In natural language processing, the field of argumentation mining aims to develop computational tools that automatically analyze and evaluate argumentation. Such tools have many uses, from automatically grading essays to identifying fallacies. In order to build such tools, annotated data is essential both for training and evaluating, especially with large language models (LLMs). Creating annotated datasets, however, presents significant challenges, not only because of the complexity of argumentation but also methodological questions such as how to represent argumentation and how to evaluate annotation quality. To create more resources as well as investigate these challenges, in this thesis, I explore several approaches to argumentation annotation. To this end, I also present a comprehensive survey of argumentation annotation. Three annotation approaches of varying complexity are explored: argumentation schemes applied to editorials, argumentative spans to online forums and political debates, and attitude annotation to tweets. The datasets thus represent a wide variety of genres and approaches. Attitude in tweets was found to show the highest agreement among annotators, while annotation of editorials with argumentation schemes was the most challenging. In the evaluation of the annotations, several types of disagreement were identified. Most saliently, disagreement often occurred in cases where multiple interpretations are possible, challenging agreement as the primary measure of quality. These findings demonstrate the need for more comprehensive evaluation approaches. I therefore demonstrate ways to evaluate beyond single agreement measures: agreement analysis from multiple angles, annotator pattern investigation, and manual inspection of disagreement. To further explore argumentation annotation, I investigate how two different LLMs annotate argumentation compared to human annotators, finding that while the models exhibit similar annotation behavior as humans, with similar agreement levels and disagreement patterns, the models agree more among themselves than human annotators.sv
dc.gup.defencedate2025-10-07
dc.gup.defenceplaceTisdagen den 7 oktober 2025, kl.13:15 i hörsal J330, Humanisten, Renströmsgatan 6sv
dc.gup.departmentDepartment of Swedish, Multilingualism, Language Technology ; Institutionen för svenska, flerspråkighet och språkteknologisv
dc.gup.dissdb-fakultetHF
dc.gup.originGöteborgs universitet. Humanistiska fakultetenswe
dc.gup.originUniversity of Gothenburg. Faculty of Humanitieseng
dc.identifier.isbn978-91-8115-302-6 (tryckt)
dc.identifier.isbn978-91-8115-303-3 (PDF)
dc.identifier.urihttps://hdl.handle.net/2077/87769
dc.language.isoengsv
dc.relation.haspart1. Anna Lindahl & Lars Borin. 2024. Annotation for computational argumentation analysis: issues and perspectives. Language and Linguistics Compass 18(1). e12505. DOI: https://doi.org/10.1111/lnc3.12505sv
dc.relation.haspart2. Anna Lindahl et al. 2019. Towards assessing argumentation annotation - a first step. In Proceedings of the 6th Workshop on Argument Mining, 177–186. Florence: ACL. DOI: 10.18653/v1/W19-4520sv
dc.relation.haspart3. Anna Lindahl. 2020. Annotating argumentation in Swedish social media. In Proceedings of the 7th Workshop on Argument Mining, 100–105. Online: ACL URL: https://aclanthology.org/2020.argmining-1.11/sv
dc.relation.haspart4. Anna Lindahl. 2025a. Annotating attitude in Swedish political tweets. In Špela Arhar Holdt et al. (eds.), Proceedings of the third workshop on resources and representations for under-resourced languages and domains (resourceful-2025), 106–110. Tallinn: University of Tartu Library URL: https://aclanthology.org/2025.resourceful-1.24/sv
dc.relation.haspart5. Anna Lindahl. 2024. Disagreement in argumentation annotation. In Gavin Abercrombie et al. (eds.), Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives), 56–66. Turin: ELRA & ICCL URL: https://aclanthology.org/2024.nlperspectives-1.6/sv
dc.relation.haspart6. Anna Lindahl. 2022. Do machines dream of artificial agreement? In Harry Bunt (ed.), Proceedings of the 18th Joint ACL - ISO Workshop on Interoperable Semantic Annotation within LREC2022, 71–75. Marseille: European Language Resources Association URL: https://aclanthology.org/2022.isa-1.9/sv
dc.relation.haspart7. Anna Lindahl. 2025b. LLMs as annotators of argumentationsv
dc.relation.ispartofseriesData linguistica 33sv
dc.subjectnatural language processingsv
dc.subjectargumentationsv
dc.subjectannotationsv
dc.subjectargumentation miningsv
dc.subjectannotation evaluationsv
dc.subjectlarge language modelssv
dc.subjectmachine learningsv
dc.titleArgumentation and agreement: Annotating and evaluating Swedish corpora for argumentation miningsv
dc.typeText
dc.type.degreeDoctor of Philosophysv
dc.type.svepDoctoral thesiseng

Files

Original bundle

Now showing 1 - 3 of 3
No Thumbnail Available
Name:
abstract.pdf
Size:
80.19 KB
Format:
Adobe Portable Document Format
Description:
abstract
No Thumbnail Available
Name:
cover.pdf
Size:
514.16 KB
Format:
Adobe Portable Document Format
Description:
cover
No Thumbnail Available
Name:
thesis.pdf
Size:
1.6 MB
Format:
Adobe Portable Document Format
Description:
Thesis

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
4.68 KB
Format:
Item-specific license agreed upon to submission
Description: