Cited By
View all- Hossain MAkhter SMahbub NHong CHuh E(2025)Why logit distillation works: A novel knowledge distillation technique by deriving target augmentation and logits distortionInformation Processing & Management10.1016/j.ipm.2024.10405662:3(104056)Online publication date: May-2025
- Mao XJiang JZeng YPeng YZhang SLi F(2024)Generative named entity recognition framework for Chinese legal domainPeerJ Computer Science10.7717/peerj-cs.242810(e2428)Online publication date: 4-Nov-2024
- Sheng YZhang ZTang PHuang BXiao Y(2024)EDAW: Enhanced Knowledge Distillation and Adaptive Pseudo Label Weights for Continual Named Entity Recognition2024 IEEE International Conference on Systems, Man, and Cybernetics (SMC)10.1109/SMC54092.2024.10831235(958-965)Online publication date: 6-Oct-2024
- Show More Cited By