You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Excuse the naiveness, I am having a bit of trouble understanding how the word-level bi-directional lstm (before the CRF layer) in the paper is reflected in the code, to be specific see the arrows in the image below.
Why is there only one word_lstm in the lm_lstm_crf.py ?
I see that in train_wc.py , if co_train is enabled, a bi-directional word level lstm is applied, but this after the CRF layer ?, shouldn't it be before the concatenated output passed to CRF as shown in the model architecture in the paper?
May be I am missing something really big & obvious :/
Thanks,
The text was updated successfully, but these errors were encountered:
By selecting co-train you just update the loss coming from LM, the bi-directional world level has been already applied on ner-model in lm_lstm_crf.py from following line in the code:
Hi there,
Excuse the naiveness, I am having a bit of trouble understanding how the word-level bi-directional lstm (before the CRF layer) in the paper is reflected in the code, to be specific see the arrows in the image below.
Why is there only one
word_lstm
in thelm_lstm_crf.py
?I see that in
train_wc.py
, ifco_train
is enabled, a bi-directional word level lstm is applied, but this after the CRF layer ?, shouldn't it be before the concatenated output passed to CRF as shown in the model architecture in the paper?May be I am missing something really big & obvious :/
Thanks,
The text was updated successfully, but these errors were encountered: