Distilling Task-Specific Knowledge from Teacher Model into BiLSTM
nlp
pytorch
transformer
korean
ele
knowledge-distillation
bert
korean-nlp
electra
distillation
bilstm
roberta
gpt-2
huggingface
t5
xlnet
distilbert
kobert
bert-multilingua
-
Updated
Jan 4, 2023 - Python