Enhancing implicit discourse relation recognition by exploiting label inter-relations
dc.contributor.advisor
Webber, Bonnie
dc.contributor.advisor
Narayanaswamy, Siddharth
dc.contributor.author
Long, Wanqiu
dc.contributor.sponsor
School of Philosophy, Psychology and Language Sciences (PPLS), University of Edinburgh
en
dc.contributor.sponsor
School of Informatics, University of Edinburgh
en
dc.date.accessioned
2025-11-10T10:23:20Z
dc.date.available
2025-11-10T10:23:20Z
dc.date.issued
2025-11-10
dc.description.abstract
Implicit Discourse Relation Recognition (IDRR) is a fundamental yet challenging task
in discourse parsing, as it involves identifying rhetorical, semantic and/or pragmatic
relationships between text spans in the absence of explicit connectives such as “because” or “however”. While recent advances leveraging pre-trained language models
and prompt-based learning have improved performance, most existing approaches treat
discourse sense labels as flat and independent categories. This neglects the rich structural information embedded in annotation frameworks like the Penn Discourse Treebank (PDTB), where discourse relations are organized hierarchically and can co-occur in some ways.
The central claim of this thesis is that structured groupings of discourse senses
— as encoded in the sense hierarchy — can serve as an effective structural prior to
guide model training, particularly by shaping how label distances are represented and
learned. This thesis proposes methods to enhance IDRR by focusing on two kinds
of label inter-relations: the hierarchical relations and co-occurrence-based label interrelations. First, we introduce a contrastive learning framework that utilizes the PDTB
sense hierarchy to guide the selection of semantically meaningful negative examples
during training, thereby encouraging the model to learn finer-grained distinctions between closely related senses. Second, we integrate hierarchical information into a
prompt-based learning paradigm through a prototype-based verbalizer, which aligns
label representations with the sense hierarchy. This approach is further extended to
support zero-shot cross-lingual IDRR, demonstrating effectiveness across both monolingual and cross-lingual scenarios. Third, we explore multi-label classification frameworks to handle cases where multiple discourse relations simultaneously hold between a single pair of text spans — an under-addressed yet prevalent phenomenon in real-world discourse. Incorporating hierarchical sense information also improves the accuracy of multi-label predictions.
Extensive experiments demonstrate the effectiveness of our approaches that consider
the label inter-relations. The results show that explicitly modeling label hierarchies
improves model performance in both single-label classification and multi-label
classification scenarios. This work advances our understanding of how structural relationships between discourse relations can be effectively utilized in computational
models, while also highlighting the importance of handling multi-label cases in discourse
relation recognition.
Finally, this thesis outlines several promising directions for future work. One avenue
is to extend the proposed approaches to broader datasets, including discourse
annotations from alternative frameworks such as RST and eRST, as well as texts from
diverse domains and languages, to better assess the generalizability of the methods.
Another direction involves integrating argument span detection with discourse relation
recognition into a unified framework, thereby advancing toward more realistic and
end-to-end discourse parsing systems. Additionally, future work may explore guiding
Large Language Models (LLMs) to better represent discourse relations and understand
the label relationships between senses by leveraging the hierarchical organization of
discourse senses. Together, these directions point to the broader goal of capturing the
complexity of coherence more effectively in natural language.
en
dc.identifier.uri
https://hdl.handle.net/1842/44149
dc.identifier.uri
http://dx.doi.org/10.7488/era/6673
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.relation.hasversion
Long, W., Narayanaswamy, S., and Webber, B. (2024). Multi-Label Classification for Implicit Discourse Relation Recognition. In Ku, L.-W., Martins, A., and Srikumar, V., editors, Findings of the Association for Computational Linguistics: ACL 2024, pages 8437–8451, Bangkok, Thailand. Association for Computational Linguistics.
en
dc.relation.hasversion
Long, W. and Webber, B. (2022). Facilitating Contrastive Learning of Discourse Relational Senses by Exploiting the Hierarchy of Sense Relations. In Goldberg, Y., Kozareva, Z., and Zhang, Y., editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 10704–10716, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics
en
dc.relation.hasversion
Long, W. and Webber, B. (2024). Leveraging Hierarchical Prototypes as the Verbalizer for Implicit Discourse Relation Recognition. arXiv, abs/2411.14880.
en
dc.relation.hasversion
Long, W., Webber, B., and Xiong, D. (2020). TED-CDB: A Large-Scale Chinese Discourse Relation Dataset on TED Talks. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2793–2803, Online. Association for Computational Linguistics.
en
dc.subject
discourse relations
en
dc.subject
sentence structure
en
dc.subject
Implicit Discourse Relation Recognition
en
dc.subject
context-aware language technology
en
dc.subject
natural language processing
en
dc.title
Enhancing implicit discourse relation recognition by exploiting label inter-relations
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en
Files
Original bundle
1 - 1 of 1
- Name:
- Long2025.pdf
- Size:
- 4.34 MB
- Format:
- Adobe Portable Document Format
- Description:
This item appears in the following Collection(s)

