Results 301 to 310 of about 4,930,132 (374)

Self-Distillation of Hidden Layers for Self-Supervised Representation Learning

open access: green
Scott Lowe   +4 more
openalex   +1 more source

Intra-class progressive and adaptive self-distillation

Neural Networks
In recent years, knowledge distillation (KD) has become widely used in compressing models, training compact and efficient students to reduce computational load and training time due to the increasing parameters in deep neural networks. To minimize training costs, self-distillation has been proposed, with methods like offline-KD and online-KD requiring ...
Jianping Gou   +5 more
openaire   +3 more sources

Home - About - Disclaimer - Privacy