Liu, X., Huang, M., Rusnachenko, N., Ive, J., Chang, J. and Zhang, J. J., 2025. Enhancing medical dialogue summarization: A MediExtract Distillation Framework. In: Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE, 6466-6473.
Full text available as:
Preview |
PDF
medi.pdf - Accepted Version Available under License Creative Commons Attribution Non-commercial. 484kB |
|
Copyright to original material in this document is with the original owner(s). Access to this content through BURO is granted on condition that you use it only for research, scholarly or other non-commercial purposes. If you wish to use it for any other purposes, you must contact BU via BURO@bournemouth.ac.uk. Any third party copyright material in this document remains the property of its respective owner(s). BU grants no licence for further use of that third party material. |
DOI: 10.1109/BIBM62325.2024.10822640
Abstract
Automatic summarization of medical dialogues, which converts colloquial doctor-patient conversations into concise notes, is increasingly important due to the growing complexity of healthcare data. However, the complexity of medical language and the lack of annotated datasets pose challenges for summarization models. In this paper, we propose a MediExtract Distillation Framework (MEDF), a novel hybrid teacher-student distillation process that leverages the power of Large Language Models (LLMs) in information capturing to enhance the performance of a smaller student model. Utilizing medical key information generated by GPT-3.5-Turbo, the model training involves two feedforward branches per iteration: one using ground truth as labels and another using generated structured medical key information as an auxiliary supervision. We validated our method on the MTS-Dialogue dataset, achieving a +2.1% improvement in BLEURT compared to previous methods, demonstrating its effectiveness in summarizing medical dialogues. Additionally, using UMLS-based BERTScore, we observed a +1.8% increase in MedBERTScore for medical term extraction, highlighting our model's practical benefits in clinical information processing. Our framework is publicly available at: https://github.com/Xiaoxiao-Liu/distill-d2n.git
| Item Type: | Book Section |
|---|---|
| ISBN: | 979-8-3503-8622-6 |
| Uncontrolled Keywords: | Hybrid Distillation: Medical Dialogue Summarization |
| Group: | Faculty of Media, Science and Technology |
| ID Code: | 41502 |
| Deposited By: | Symplectic RT2 |
| Deposited On: | 10 Mar 2026 16:28 |
| Last Modified: | 10 Mar 2026 16:28 |
Downloads
Downloads per month over past year
| Repository Staff Only - |
Tools
Tools