• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Konferenzschrift
  4. Exploring the Limits of Model Compression in LLMs: A Knowledge Distillation Study on QA Tasks
 
  • Details
  • Full
Options
August 2025
Conference Paper
Title

Exploring the Limits of Model Compression in LLMs: A Knowledge Distillation Study on QA Tasks

Abstract
Large Language Models (LLMs) have demonstrated outstanding performance across a range of NLP tasks; however, their computational demands hinder their deployment in real-world, resource-constrained environments. This work investigates the extent to which LLMs can be compressed using Knowledge Distillation (KD) while maintaining strong performance on Question Answering (QA) tasks. We evaluate student models distilled from the Pythia and Qwen2.5 families on two QA benchmarks, SQuAD and MLQA, under zero-shot and oneshot prompting conditions. Results show that student models retain over 90% of their teacher models' performance while reducing parameter counts by up to 57.1%. Furthermore, oneshot prompting yields additional performance gains over zero-shot setups for both model families. These findings underscore the trade-off between model efficiency and task performance, demonstrating that KD, combined with minimal prompting, can yield compact yet capable QA systems suitable for resource-constrained applications.
Author(s)
Datta, Joyeeta
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Doll, Niclas
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Ramadan, Qusai
Boukhers, Zeyd  
Fraunhofer-Institut für Angewandte Informationstechnik FIT  
Mainwork
SIGDIAL 2025, 26th Annual Meeting of the Special Interest Group on Discourse and Dialogue. Proceedings  
Conference
Special Interest Group on Discourse and Dialogue (SIGdial Annual Meeting) 2025  
Open Access
File(s)
Download (374.51 KB)
Rights
CC BY-SA 4.0: Creative Commons Attribution-ShareAlike
DOI
10.18653/v1/2025.sigdial-1.39
10.24406/publica-5756
Language
English
Fraunhofer-Institut für Angewandte Informationstechnik FIT  
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Keyword(s)
  • Knowledge Distillation

  • Large Language Model

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024