#Model Distillation
Latest news articles tagged with "Model Distillation". Follow the timeline of events, related topics, and entities.
Articles (1)
-
πΊπΈ Fast and Effective On-policy Distillation from Reasoning Prefixes
[USA]
arXiv:2602.15260v1 Announce Type: cross Abstract: On-policy distillation (OPD), which samples trajectories from the student model and supervises them with a teacher at the token level, avoids relying...
Related: #Machine Learning, #Natural Language Processing, #Reinforcement Learning