Bidirectional encoder representations from transformers (bert) Encoding positional transformer embedding attention bert nlp harvard annotated encoder transformers
Sine Position Embedding
Bert embedding position desirable positional properties pe sine follows dot wave vectors between case two What are the desirable properties for positional embedding in bert