You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, dear ziyi~
I found in your code, the bert output weights are not set to be the same as the input embedding, which can be proved in here(In detail, the code didnt set the weight of BertLMPredictionHead.decoder to be the same as the weight of BertModel.embeddings).
I think maybe it is you who deliberately modified the source code of bert LM into this way. Why? Will it influence the final result?
The text was updated successfully, but these errors were encountered:
Hi, yes the input and output embedding layers are not shared because I didn't define get_input/output_embeddings in BERT. This was not intentional but I found this has almost no impact on the alignment performance.
Hi, dear ziyi~
I found in your code, the bert output weights are not set to be the same as the input embedding, which can be proved in here(In detail, the code didnt set the weight of
BertLMPredictionHead.decoder
to be the same as the weight ofBertModel.embeddings
).I think maybe it is you who deliberately modified the source code of bert LM into this way. Why? Will it influence the final result?
The text was updated successfully, but these errors were encountered: