I. Introduction
As artificial intelligence continues to evolve rapidly, the role of large language models in natural language processing has become increasingly prominent [
1]. Their ability to comprehend and generate human language with high accuracy makes them foundational in a wide range of applications. Despite their effectiveness, LLMs typically involve vast numbers of parameters and demand significant computational power, which hinders their practical deployment [
2]. To address this, knowledge distillation has gained attention as an effective approach. It compresses large models by transferring their knowledge to smaller ones, thereby improving efficiency without compromising performance and enabling more feasible deployment in real-world scenarios.
Traditional knowledge distillation methods typically rely on a single-teacher model to guide the student model. This is achieved by minimizing the difference between the outputs of the teacher and the student, allowing the student to learn the knowledge structure of the teacher. However, in complex tasks or diverse data distributions, a single teacher often fails to cover all key knowledge areas. This limitation restricts the performance of the student model. In such cases, multi-teacher distillation strategies have gained attention. These approaches aim to incorporate multiple teachers with different characteristics to provide richer, more diverse, and complementary knowledge, thus improving the student's overall learning capacity and generalization ability [
3].
Multi-teacher distillation not only broadens the scope and depth of knowledge transfer but also raises important questions about teacher coordination, knowledge fusion strategies, and conflict resolution. In a multi-teacher setting, different teacher models may produce varying outputs for the same input. These differences may reflect redundant or conflicting information. A key challenge is to reconcile these outputs, explore their complementarity, and convert them into effective supervisory signals for the student [
4,
5]. This requires reasonable integration of the teacher outputs and dynamic adjustment of the student model's learning focus to achieve optimal distillation outcomes.
Moreover, large language models possess strong semantic abstraction and context awareness. Their rich knowledge representations demand higher requirements during the distillation process. Unlike traditional classification or small sequence models, language models present knowledge that is hierarchical and context-dependent. Therefore, it is not sufficient to focus only on output consistency. It is essential to consider internal feature representations, attention mechanisms, and information transfer across multiple layers. Under multi-teacher guidance, these complex signals can be captured from different perspectives, offering a more comprehensive learning framework for the student and promoting a deeper understanding of language knowledge.
II. Related Work and Foundation
Knowledge distillation has emerged as a vital methodology for compressing large language models (LLMs), enabling their deployment in resource-constrained environments without significant loss of performance. Early distillation approaches typically leveraged a single teacher to guide the student model by minimizing the difference in their output distributions. However, recent advances have extended this paradigm by exploring multi-granularity and multi-teacher strategies to capture richer and more diverse knowledge. Liu et al. proposed multi-granularity structural knowledge distillation that aligns not only output logits but also intermediate features, significantly improving language model compression effectiveness [
6]. Similarly, Li et al. developed dynamic knowledge distillation approaches for pre-trained language models, introducing adaptive teacher-student coordination mechanisms for more effective transfer [
7]. To further increase compression efficiency, Kai et al. combined distillation with fine-tuning, demonstrating that carefully designed distillation processes yield student models with strong generalization capabilities and practical size reduction [
8].
The increasing complexity of language representations in LLMs has highlighted the importance of transferring not just output-level knowledge, but also internal semantic structures and contextual features. Structured gradient guidance [
9] and low-rank adaptation with semantic supervision [
10] have both been shown to improve few-shot adaptation and generalization in large models by ensuring semantic consistency during fine-tuning. Knowledge graph reasoning has also been integrated with pre-trained models to support structured anomaly detection, emphasizing the value of external structured knowledge in semantic tasks [
11]. Advances in structured memory mechanisms [
12] and structured knowledge integration [
13] further enhance context representation stability and long-term memory modeling in LLMs, which are particularly beneficial for knowledge-intensive and multi-turn tasks.
Addressing the limitations of single-teacher guidance, multi-source fusion strategies have been introduced to facilitate the integration of diverse perspectives. Wang explored time-aware and multi-source feature fusion for transformer models, which serves as a foundation for multi-teacher distillation by leveraging complementary knowledge from heterogeneous sources [
14]. In parallel, the use of structured preference modeling [
15] and graph-based spectral decomposition [
16] for fine-tuning illustrates how reinforcement learning and graph-theoretic approaches can effectively coordinate model parameters and optimize adaptation for downstream tasks.
Another significant direction in LLM research is the enhancement of generation quality and context management. He et al. proposed context-guided dynamic retrieval for improving generation quality in retrieval-augmented generation (RAG) models [
17], while Peng tackled hallucination detection through context-aligned and evidence-based verification mechanisms [
18]. Deng focused on transfer methods suitable for low-resource scenarios, revealing that sophisticated transfer and adaptation strategies are necessary to maintain generation quality and coherence in challenging environments [
19].
Several works also highlight the application of LLMs to specialized semantic tasks and system-level challenges. Wang demonstrated the use of LLama-based modeling for semantic detection in social media security applications [
20], while Guo et al. introduced perception-guided frameworks for model design tailored to complex language understanding tasks [
21]. Moreover, Fang provided a predictive framework for backend latency using AI-augmented structured modeling, underscoring the importance of efficiency and response speed in practical deployments [
22].
Together, these studies reflect the field’s progression from single-teacher, output-only distillation toward multi-teacher, structure-aware, and semantically guided knowledge transfer. This evolution not only enhances student model performance and generalization, but also establishes a technical foundation for deploying highly efficient, adaptive, and robust language models in diverse application domains.
III. Method
This study aims to construct a large language model distillation strategy under the collaborative guidance of multiple teachers to improve the language understanding and generation capabilities of the student model while maintaining a small parameter scale. The model architecture is shown in
Figure 1.
First, assume that there are multiple teacher models
and a student model S to be trained. Each teacher model generates a corresponding output probability distribution
on the same input x, and the output of the student model is
. To integrate the knowledge of multiple teachers, we first define a unified teacher target distribution as a weighted average form:
Among them, is the weight coefficient of the teacher model, which satisfies . This fusion mechanism can dynamically integrate the knowledge of each teacher during the distillation process and enhance the diversity of learning signals of the student model.
In order to guide the student model to learn the output behavior of the teacher model, we adopt Kullback-Leibler divergence as the main distillation loss function. Specifically, the distillation loss is defined as:
Where represents the true label distribution of the samples.
During the joint training process, we perform a weighted combination of the above two losses to form the final training objective function:
Among them,
controls the trade-off between knowledge distillation and supervised learning. In addition, to address the possible information conflicts between multiple teachers, we designed a dynamic weight adjustment mechanism based on information entropy, so that the influence of each teacher model is adjusted according to the uncertainty of its output. Specifically, the weight of the teacher
is defined as:
The information entropy is . The smaller the entropy value, the more certain the output of the teacher model is, and the higher its weight is, which is conducive to improving the learning efficiency of the student model for explicit knowledge.
In order to further improve the distillation effect, we also introduced an intermediate feature-matching strategy so that the student model not only focuses on the consistency of the final output during training but also imitates the representation of the intermediate layer of the teacher model. Assuming that the feature representations of the student model and the teacher model in the intermediate layer are
and
respectively, the feature distillation loss can be defined as:
Where is the importance weight of each teacher's intermediate feature. Finally, all loss terms can be integrated into a multi-objective optimization framework to comprehensively improve the comprehensive performance of the student model in terms of expression ability, generalization ability, and learning efficiency.
VI. Experimental Results
A. Dataset
This study uses the C4 (Colossal Clean Crawled Corpus) dataset as the primary corpus for both training and distillation. C4 is an open-source dataset built from large-scale web text. It has undergone strict data cleaning and content filtering to remove low-quality, duplicated, and non-natural language content. As a result, it offers high linguistic quality and rich semantic information. The dataset covers a wide range of language use cases, including encyclopedias, news articles, forums, and technical documents. This diversity supports the development of general-purpose language understanding models.
The C4 dataset is several hundred gigabytes in size and contains billions of text segments. It provides large language models with sufficiently diverse training samples. In this study, the English subset of C4 is selected. It preserves coherent contextual structures and allows the model to generalize across different writing styles and domains. This data foundation helps evaluate the performance and adaptability of the multi-teacher distillation strategy under real-world corpus conditions.
In addition, the C4 dataset has been widely used in pretraining and distillation research for large language models. Its text length and distribution characteristics closely match real-world applications. It serves as a suitable intermediate resource for student models to learn the language capabilities of teacher models during distillation. By using this dataset, the transfer of multi-level semantic features and language expression abilities can be effectively achieved. This provides a strong data basis for the deployment and practical use of distilled models.
B. Experimental Results
To begin the empirical analysis, this paper first designs and implements a comparative experiment aimed at evaluating the performance of the proposed method against several representative baseline approaches. The primary objective of this comparison is to establish a clear benchmark for assessing the effectiveness of the multi-teacher collaborative distillation strategy. In this setting, multiple widely adopted distillation models are selected to provide a reference framework for evaluation. The comparison spans multiple core dimensions of model performance, ensuring a comprehensive examination of the proposed technique. The detailed outcomes and relevant metrics of this comparative study are systematically organized and presented in
Table 1 for clarity and ease of analysis.
As shown in the table, the proposed multi-teacher collaborative distillation strategy consistently outperforms mainstream methods across key metrics, achieving notable gains in Perplexity, Distillation Loss, and BLEU. The student model’s perplexity improves from 24.7 to 20.8, indicating enhanced contextual prediction accuracy, while the distillation loss drops to 1.64—lower than MiniLM (2.18) and DKD (1.97)—demonstrating more stable and efficient knowledge transfer. The integration of multiple teachers provides richer, complementary semantic representations, reducing noise and improving alignment. Additionally, the method achieves the highest BLEU score of 86.7, reflecting superior text generation quality and semantic completeness. To further explore the influence of teacher diversity, the study examines the effect of varying the number of teacher models on distillation performance, with results summarized in
Figure 2.
As shown in
Figure 2, the performance of the student model improves consistently across multiple evaluation metrics as the number of teacher models increases. The Perplexity value decreases from approximately 25.4 with a single teacher to 20.8 with five teachers. This indicates a significant enhancement in the student model's ability to model language distributions. The lower perplexity suggests that multi-teacher guidance helps the student capture richer contextual information.
For the Distillation Loss metric, the KL divergence gradually declines from 2.42 to 1.64 as the number of teachers increases from one to five. This shows that the student model fits the knowledge distributions of multiple teachers more accurately. It confirms the effectiveness of the multi-teacher fusion mechanism in the knowledge transfer process. The mechanism supports comprehensive semantic transmission and reduces the representational gap between teacher and student models.
The change in the BLEU score further supports this finding. It increases from around 79.1 to 86.7, showing that multi-teacher distillation not only enhances language understanding but also improves the quality of language generation. The diversity in expression styles and semantic structures provided by different teachers offers a broader linguistic reference for the student model.
This paper further gives the collaborative distillation performance evaluation in the multi-task learning scenario, and the experimental results are shown in
Figure 3.
Figure 3 illustrates the performance of the proposed multi-teacher collaborative distillation method in a multi-task learning setting. Overall, the method achieves high accuracy across five representative natural language processing tasks. This indicates that the proposed strategy has strong cross-task generalization ability. It can effectively adapt to both language understanding and generation tasks.
In specific tasks, the model achieves 88.7% accuracy in Summarization and 87.3% in Paraphrase. This shows that collaborative distillation helps the student model capture core semantics in long texts. It also supports high-quality language reconstruction while maintaining consistency in expression. These improvements come from the complementarity among teacher models in summarization and semantic transformation, which broadens the expressive capacity of the student.
For structured understanding tasks such as NER and QA, the model also performs consistently, with accuracies of 86.1% and 85.2%, respectively. This suggests that the multi-level linguistic knowledge and contextual modeling provided by multiple teachers support more accurate sequence labeling and question-answer matching. These results show that collaborative distillation enhances both local token-level perception and global semantic understanding in the student model.
Although the accuracy for the Sentiment task is slightly lower at 83.5%, it remains at a high level. This may be due to the task's reliance on fine-grained semantic features for sentiment inference. While there is room for further improvement in this area, the overall results confirm the generality and effectiveness of the collaborative distillation method within a multi-task learning framework. It provides a feasible path toward building lightweight and high-performing multifunctional language models.
V. Conclusion
This study conducts an in-depth investigation into the application of multi-teacher collaborative distillation for compressing and optimizing large language models. It systematically proposes a distillation framework that integrates knowledge from multiple sources to address the limitations of traditional single-teacher methods in knowledge coverage and representation capacity. By introducing multiple teacher models and designing a unified output fusion mechanism along with a feature alignment strategy, the student model can acquire more complete and diverse linguistic knowledge while maintaining a compact size. Experimental results demonstrate that the proposed method achieves significant improvements across multiple evaluation metrics, effectively enhancing the overall performance of the student model in both language modeling and generation tasks.
In a multi-teacher setting, managing differences in knowledge, weight assignment, and output conflicts among teachers is a central challenge in the distillation process. This study addresses the challenge through an entropy-guided dynamic weighting mechanism and intermediate layer feature matching. These strategies enable an optimized combination of contributions from different teachers. This approach improves not only the efficiency of model compression but also the generalization and task adaptability of the student model. It provides a promising technical pathway for deploying large models in practical applications. Further validation in multi-task scenarios confirms the generality and scalability of the proposed method.
This research contributes positively to the field of natural language processing applications. In real-world scenarios such as dialogue systems, text generation, intelligent customer service, question answering, and mobile deployment, models are required to meet increasingly strict demands for computational efficiency and response speed. The balance of performance and efficiency achieved through multi-teacher collaborative distillation allows small models to perform at a level close to or even surpassing that of large models while maintaining runtime efficiency. This capability is of great importance for advancing intelligent language technologies toward lightweight, real-time, and edge-computing-friendly solutions. Looking ahead, there remain several directions for advancing collaborative distillation. These include modeling and resolving knowledge conflicts among teacher models, optimizing collaborative structures for cross-lingual or cross-modal tasks, and integrating distillation with reinforcement learning or self-supervised learning techniques. In addition, building distillation frameworks with incremental learning capabilities aligned with the evolving nature of large language models will be a key area of development. Continued exploration in these directions will lay a solid foundation for creating smarter, more efficient, and more deployable language model ecosystems.
References
- Y. Gu, L. Dong, F. Wei, et al., "MiniLLM: knowledge distillation of large language models," arXiv preprint, arXiv:2306.08543, 2023.
- X. Xu, M. Li, C. Tao, et al., "A survey on knowledge distillation of large language models," arXiv preprint, arXiv:2402.13116, 2024.
- C. Yang, Y. Zhu, W. Lu, et al., "Survey on knowledge distillation for large language models: methods, evaluation, and application," ACM Transactions on Intelligent Systems and Technology, 2024.
- S. Muralidharan, S. T. Sreenivas, R. Joshi, et al., "Compact language models via pruning and knowledge distillation," Advances in Neural Information Processing Systems, vol. 37, pp. 41076–41102, 2024.
- J. Liu, C. Zhang, J. Guo, et al., "DDK: distilling domain knowledge for efficient large language models," Advances in Neural Information Processing Systems, vol. 37, pp. 98297–98319, 2024.
- C. Liu, C. Tao, J. Feng, et al., "Multi granularity structural knowledge distillation for language model compression," Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1001–1011, 2022.
- L. Li, Y. Lin, S. Ren, et al., "Dynamic knowledge distillation for pre trained language models," arXiv preprint, arXiv:2109.11295, 2021.
- A. Kai, L. Zhu, and J. Gong, "Efficient compression of large language models with distillation and fine tuning," Journal of Computer Science and Software Applications, vol. 3, no. 4, pp. 30–38, 2023.
- H. Zheng, Y. Wang, R. Pan, G. Liu, B. Zhu, and H. Zhang, "Structured gradient guidance for few-shot adaptation in large language models," arXiv preprint, arXiv:2506.00726, 2025.
- H. Zheng, Y. Ma, Y. Wang, G. Liu, Z. Qi, and X. Yan, "Structuring low-rank adaptation with semantic guidance for model fine-tuning," 2025.
- X. Liu, Y. Qin, Q. Xu, Z. Liu, X. Guo, and W. Xu, "Integrating knowledge graph reasoning with pretrained language models for structured anomaly detection," 2025.
- Y. Xing, T. Yang, Y. Qi, M. Wei, Y. Cheng, and H. Xin, "Structured memory mechanisms for stable context representation in large language models," arXiv preprint, arXiv:2505.22921, 2025.
- Y. Peng, "Structured knowledge integration and memory modeling in large language systems," Transactions on Computational and Scientific Methods, vol. 4, no. 10, 2024.
- X. Wang, "Time-aware and multi-source feature fusion for transformer-based medical text analysis," Transactions on Computational and Scientific Methods, vol. 4, no. 7, 2024.
- L. Zhu, F. Guo, G. Cai, and Y. Ma, "Structured preference modeling for reinforcement learning-based fine-tuning of large models," Journal of Computer Technology and Software, vol. 4, no. 4, 2025.
- H. Zhang, Y. Ma, S. Wang, G. Liu, and B. Zhu, "Graph-based spectral decomposition for parameter coordination in language model fine-tuning," arXiv preprint, arXiv:2504.19583, 2025.
- J. He, G. Liu, B. Zhu, H. Zhang, H. Zheng, and X. Wang, "Context-guided dynamic retrieval for improving generation quality in RAG models," arXiv preprint, arXiv:2504.19436, 2025.
- Y. Peng, "Context-aligned and evidence-based detection of hallucinations in large language model outputs," Transactions on Computational and Scientific Methods, vol. 5, no. 6, 2025.
- Y. Deng, "Transfer methods for large language models in low-resource text generation tasks," Journal of Computer Science and Software Applications, vol. 4, no. 6, 2024.
- R. Wang, "Joint semantic detection and dissemination control of phishing attacks on social media via LLama-based modeling," arXiv preprint, arXiv:2504.00282, 2025.
- F. Guo, L. Zhu, Y. Wang, and G. Cai, "Perception-guided structural framework for large language model design," Journal of Computer Technology and Software, vol. 4, no. 5, 2025.
- Z. Fang, "A deep learning-based predictive framework for backend latency using AI-augmented structured modeling," Journal of Computer Technology and Software, vol. 3, no. 7, 2024.
- X. Jiao, Y. Yin, L. Shang, et al., "TinyBERT: distilling BERT for natural language understanding," arXiv preprint, arXiv:1909.10351, 2019.
- Z. Sun, H. Yu, X. Song, et al., "MobileBERT: a compact task-agnostic BERT for resource-limited devices," arXiv preprint, arXiv:2004.02984, 2020.
- W. Wang, F. Wei, L. Dong, et al., "MiniLM: deep self-attention distillation for task-agnostic compression of pre-trained transformers," Advances in Neural Information Processing Systems, vol. 33, pp. 5776–5788, 2020.
- B. Zhao, Q. Cui, R. Song, et al., "Decoupled knowledge distillation," Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11953–11962, 2022.
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).