Skip to main content

nclu_team at SemEval-2023 Task 6: Attention-based Approaches for Large Court Judgement Prediction with Explanation.

Rusnachenko, N., Markchom, T. and Liang, H., 2023. nclu_team at SemEval-2023 Task 6: Attention-based Approaches for Large Court Judgement Prediction with Explanation. In: Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023). Association for Computational Linguistics (ACL), 270-274.

Full text available as:

[thumbnail of nclu.pdf]
Preview
PDF
nclu.pdf - Published Version
Available under License Creative Commons Attribution.

235kB

DOI: 10.18653/v1/2023.semeval-1.36

Abstract

Legal documents tend to be large in size. In this paper, we provide an experiment with attention-based approaches complemented by certain document processing techniques for judgment prediction. For the prediction of explanation, we consider this as an extractive text summarization problem based on an output of (1) CNN with attention mechanism and (2) self-attention of language models. Our extensive experiments show that treating document endings at first results in a 2.1% improvement in judgment prediction across all the models. Additional content peeling from non-informative sentences allows an improvement of explanation prediction performance by 4% in the case of attention-based CNN models. The best submissions achieved 8th and 3rd ranks on judgment prediction (C1) and prediction with explanation (C2) tasks respectively among 11 participating teams. The results of our experiments are published<sup>1</sup>

Item Type:Book Section
ISBN:9781959429999
Additional Information:The 17th International Workshop on Semantic Evaluation (SemEval-2023) Toronto, Canada 13–14 July 2023
Group:Faculty of Media, Science and Technology
ID Code:41507
Deposited By: Symplectic RT2
Deposited On:20 Mar 2026 16:05
Last Modified:20 Mar 2026 16:05

Downloads

Downloads per month over past year

More statistics for this item...
Repository Staff Only -