Guohao Li (@guohao_li)

GLM-5의 on-policy cross-stage distillation이 흥미롭다는 언급입니다. 작성자는 @thinkymachines의 on-policy distillation 블로그를 인용하며, 특히 옴니(omni) 모델에서 재학습 시 발생하는 catastrophic forgetting(망각)을 완화하기 위한 on-policy distillation 활용 아이디어를 공유했습니다.

https://x.com/guohao_li/status/2024071081913381337

#glm5 #distillation #continuallearning #onpolicy

Guohao Li 🐫 (@guohao_li) on X

The on-policy cross-stage distillation in GLM-5 by @Zai_org is quite interesting as well. When @thinkymachines released their blog on on-policy distillation, I also shared some thoughts on how to use on-policy distillation for catastrophic forgetting, especially for omni models.

X (formerly Twitter)