#Knowledge Distillation
Latest news articles tagged with "Knowledge Distillation". Follow the timeline of events, related topics, and entities.
Articles (5)
-
πΊπΈ Decoder-based Sense Knowledge Distillation
[USA]
arXiv:2602.22351v1 Announce Type: cross Abstract: Large language models (LLMs) learn contextual embeddings that capture rich semantic information, yet they often overlook structured lexical knowledge...
Related: #AI Research, #Natural Language Processing -
πΊπΈ Reinforcement-aware Knowledge Distillation for LLM Reasoning
[USA]
arXiv:2602.22495v1 Announce Type: cross Abstract: Reinforcement learning (RL) post-training has recently driven major gains in long chain-of-thought reasoning large language models (LLMs), but the hi...
Related: #Machine Learning, #Reinforcement Learning -
πΊπΈ STAR: Similarity-guided Teacher-Assisted Refinement for Super-Tiny Function Calling Models
[USA]
arXiv:2602.03022v2 Announce Type: replace Abstract: The proliferation of Large Language Models (LLMs) in function calling is pivotal for creating advanced AI agents, yet their large scale hinders wid...
Related: #AI Model Optimization, #Function Calling -
πΊπΈ Trust the uncertain teacher: distilling dark knowledge via calibrated uncertainty
[USA]
arXiv:2602.12687v1 Announce Type: cross Abstract: The core of knowledge distillation lies in transferring the teacher's rich 'dark knowledge'-subtle probabilistic patterns that reveal how classes are...
Related: #Model Compression, #Uncertainty Quantification -
πΊπΈ FastWhisper: Adaptive Self-knowledge Distillation for Real-time Automatic Speech Recognition
[USA]
arXiv:2601.19919v1 Announce Type: cross Abstract: Knowledge distillation is one of the most effective methods for model compression. Previous studies have focused on the student model effectively tra...
Related: #Artificial Intelligence, #Model Compression