sequence tagging for NER for ULMFiT
to replicate result:
you can download the data/
folder from here, and put it in root directory.
I am currently doing experiments in jupyter notebook coNLL_three_layer.ipynb
-
concat both forward and backward outputs from language model
W_LM = [W_forward, W_backward]
-
feeding word vectors from GloVe to a BiLSTM and get output
W_glove
-
concatenating these outputs
W = [W_glove, W_LM]
-
feeding
W
to another BiLSTM to get final result.
F1 score of 76.
(need to improve by fine tuning parameters, see how the toks are preprocessed, adding char embedding, adding CRF layer.
-
which layer of lanuage model should be used for Sequence tagging problem
-
how to build a better language model for sequence tagging
Regularizing and Optimizing LSTM Language Models
deep contextualized word representations
End-to-end Sequence Labeling via Bi-directional LSTM-CNNs-CRF
Semi-supervised sequence tagging with bidirectional language models