Sabaragamuwa University of Sri Lanka

Task-specific knowledge distillation for efficient natural language processing applications

Show simple item record

dc.contributor.author Balasubramaniam, G
dc.contributor.author Abishethvarman, V
dc.contributor.author Prasanth, S
dc.contributor.author Kumara, B.T.G.S.
dc.date.accessioned 2026-01-17T07:14:11Z
dc.date.available 2026-01-17T07:14:11Z
dc.date.issued 2025-12-03
dc.identifier.issn 2815-0341
dc.identifier.uri http://repo.lib.sab.ac.lk:8080/xmlui/handle/susl/5179
dc.description.abstract Large Language Models (LLMs) excel at a variety of Natural Language Processing (NLP) tasks, but their implementation is frequently hampered by high computational costs and inefficiency in low-resource settings. While smaller models provide faster inference, they often have worse contextual knowledge, greater mistake rates, and are more susceptible to hallucinations. This paper investigates task-specific knowledge distillation as a realistic method for transferring skills from high-capacity teacher models to compact student models in three important NLP tasks: text summarization, sentiment analysis, and text categorisation. Information was condensed from four powerful teacher models LLaMA 3.1 (70B), Falcon (40B), Gemma2 (10B), and Qwen2.5 (72B), into much smaller students (8B, 7B, 2B, and 7B, respectively), and their performance was assessed using conventional task-specific measures. The results reveal that distilled models preserve a significant percentage of instructor performance while minimizing hallucinations and increasing efficiency. In summary, Qwen2.5 (7B) obtained ROUGE-L of 0.6743, BLEU of 47.6483, and METEOR of 0.6726. In sentiment analysis, distillation increased LLaMA 8B accuracy from 0.4025 to 0.5900, and in classification, distilled models better caught overlapping category meanings. These findings demonstrate that task-specific distillation is a viable strategy for developing lightweight, high-performance NLP models for resource-constrained applications. en_US
dc.language.iso en en_US
dc.publisher Sabaragamuwa University of Sri Lanka en_US
dc.subject Hallucination en_US
dc.subject Knowledge distillation en_US
dc.subject Large Language Models (LLM) en_US
dc.subject Sentiment analysis en_US
dc.subject Text summarization en_US
dc.title Task-specific knowledge distillation for efficient natural language processing applications en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account